From Beating Humans to Shaping Society: Key Takeaways from the AI Index 2024 Report

The Stanford Institute for Human-Centered Artificial Intelligence recently released their AI Index 2024 Annual Report, providing a comprehensive look at the rapidly evolving AI landscape over the past year. Here are some key excerpts from the report:

Message From the Co-directors, Ray Perrault and Jack Clark:

“A decade ago, the best AI systems in the world were unable to classify objects in images at a human level. AI struggled with language comprehension and could not solve math problems. Today, AI systems routinely exceed human performance on standard benchmarks.

Progress accelerated in 2023. New state-of-the-art systems like GPT-4, Gemini, and Claude 3 are impressively multimodal: They can generate fluent text in dozens of languages, process audio, and even explain memes. As AI has improved, it has increasingly forced its way into our lives. Companies are racing to build AI-based products, and AI is increasingly being used by the general public. But current AI technology still has significant problems. It cannot reliably deal with facts, perform complex reasoning, or explain its conclusions.

AI faces two interrelated futures. First, technology continues to improve and is increasingly used, having major consequences for productivity and employment. It can be put to both good and bad uses. In the second future, the adoption of AI is constrained by the limitations of the technology. Regardless of which future unfolds, governments are increasingly concerned. They are stepping in to encourage the upside, such as funding university R&D and incentivizing private investment. Governments are also aiming to manage the potential downsides, such as impacts on employment, privacy concerns, misinformation, and intellectual property rights.

As AI rapidly evolves, the AI Index aims to help the AI community, policymakers, business leaders, journalists, and the general public navigate this complex landscape. It provides ongoing, objective snapshots tracking several key areas: technical progress in AI capabilities, the community and investments driving AI development and deployment, public opinion on current and potential future impacts, and policy measures taken to stimulate AI innovation while managing its risks and challenges. By comprehensively monitoring the AI ecosystem, the Index serves as an important resource for understanding this transformative technological force.

On the technical front, this year’s AI Index reports that the number of new large language models released worldwide in 2023 doubled over the previous year. Two-thirds were open-source, but the highest-performing models came from industry players with closed systems. Gemini Ultra became the first LLM to reach human level performance on the Massive Multitask Language Understanding (MMLU) benchmark; performance on the benchmark has improved by 15 percentage points since last year. Additionally, GPT-4 achieved an impressive 0.96 mean win rate score on the comprehensive Holistic Evaluation of Language Models (HELM) benchmark, which includes MMLU among other evaluations.

Although global private investment in AI decreased for the second consecutive year, investment in generative AI skyrocketed. More Fortune 500 earnings calls mentioned AI than ever before, and new studies show that AI tangibly boosts worker productivity. On the policymaking front, global mentions of AI in legislative proceedings have never been higher. U.S. regulators passed more AI-related regulations in 2023 than ever before. Still, many expressed concerns about AI’s ability to generate deepfakes and impact elections. The public became more aware of AI, and studies suggest that they responded with nervousness.”

The report also summarized the top 10 takeaways:

  • AI beats humans on some tasks, but not on all.
  • Industry continues to dominate frontier AI research.
  • Frontier models get way more expensive.
  • The United States leads China, the EU, and the U.K. as the leading source of top AI models.
  • Robust and standardized evaluations for LLM responsibility are seriously lacking.
  • Generative AI investment skyrockets.
  • The data is in: AI makes workers more productive and leads to higher quality work.
  • Scientific progress accelerates even further, thanks to AI.
  • The number of AI regulations in the United States sharply increases.
  • People across the globe are more cognizant of AI’s potential impact—and more nervous.

To read the full report, visit https://aiindex.stanford.edu/report/

Attribution: Content excerpted from “The AI Index 2024 Annual Report” by Nestor Maslej, Loredana Fattorini, Raymond Perrault, Vanessa Parli, Anka Reuel, Erik Brynjolfsson, John Etchemendy, Katrina Ligett, Terah Lyons, James Manyika, Juan Carlos Niebles, Yoav Shoham, Russell Wald, and Jack Clark; AI Index Steering Committee, Institute for Human-Centered AI, Stanford University. Licensed under Attribution-NoDerivatives 4.0 International.


Comments

Leave a Reply

Your email address will not be published. Required fields are marked *