The Dawn of the AI Arms Race A New Era of Technological

The Dawn of the AI Arms Race A New Era of Technological

The dawn of the AI arms race The artificial intelligence arms race, once focused on creating colossal models trained on vast datasets to replicate human-level intelligence, is undergoing a transformation. Tech giants and startups are now prioritizing the development of more compact AI software that is more affordable, faster, and tailored for specific tasks. This shift marks the beginning of a new era in The Dawn of the AI Arms Race, where the emphasis is on efficiency, specialization, and practical applications.

The Rise of Small and Medium Language Models

These smaller AI models are designed for specific tasks and are trained on smaller datasets. While large models like OpenAI’s GPT-4 cost over $100 million to develop and utilize more than one trillion parameters, smaller models can be developed for less than $10 million and use fewer than 10 billion parameters. This makes them more cost-efficient, requiring less computing power for each query.

Smaller AI models offer cost-efficiency with tailored capabilities, ideal for focused tasks with scaled budgets, acording to WSJ Print Subscription

Microsoft’s Phi Models

Microsoft has introduced its suite of small models named Phi. According to CEO Satya Nadella, these models are 1/100th the size of the free model behind OpenAI’s ChatGPT but can perform many tasks nearly as well. Yusuf Mehdi, Microsoft’s chief commercial officer, emphasized the importance of a diverse array of models for the future. The escalating costs of operating large models prompted Microsoft to pivot towards more economical options.

AI Laptops and Device Efficiency

Recently, Microsoft launched AI laptops that leverage multiple AI models for tasks like search and image generation. These models operate efficiently on the device itself, eliminating the need for access to powerful cloud-based supercomputers. Similarly, Google, Mistral, Anthropic, and Cohere have also introduced smaller models this year. Apple has revealed its AI roadmap, intending to deploy small models that run entirely on devices, enhancing speed and security.


Samsung Electronics Anticipates Significant Second-Quarter Earnings Surge

Samsung Electronics Anticipates Significant Second-Quarte

Samsung Electronics is poised for a robust recovery in the second quarter of 2024, buoyed by a revitalized semiconductor division…


OpenAI’s Cost-Effective Approach

OpenAI, a pioneer of large models, has released a more cost-effective version of its flagship model. They are open to developing smaller models in the future. Large models are excessive for many applications, such as document summarization or image generation. Illia Polosukhin, a blockchain technologist and co-author of a pivotal 2017 Google paper on generative AI, noted, “It shouldn’t take quadrillions of operations to compute 2 + 2.

The Economics of Smaller Models

Both businesses and consumers seek economical ways to utilize generative AI technologies, whose returns are still uncertain. Smaller models, requiring less computing power, can answer queries at a fraction of the cost. According to Yoav Shoham, co-founder of AI21 Labs, they can be as effective as large models. Fine-tuning these smaller models on specific data sets enhances their performance.

Industry Adoption and Efficiency

Experian transitioned from large models to small ones for their AI chatbots used in financial advice and customer service. Trained on the company’s internal data, these smaller models matched the performance of larger ones at a significantly lower cost, according to Ali Khan, Experian’s chief data officer. Clara Shih, head of AI at Salesforce, noted that smaller models offer faster performance and avoid the latency issues of large models.

A Shift in Focus

The shift towards smaller models comes as progress on publicly released large models decelerates. Since OpenAI’s release of GPT-4, no new models have achieved a similar leap. This slowdown, attributed to a shortage of high-quality new data for training, has directed attention towards more efficient, smaller models. Sébastien Bubeck, the Microsoft executive leading the Phi model project, remarked, “There is this little moment of lull where everybody is waiting. It makes sense that your attention gets diverted to, ‘OK, can you actually make this stuff more efficient?'”

The Future of AI

Whether this lull is temporary or indicative of a broader technological challenge remains uncertain. The trend towards smaller models underscores AI’s evolution from spectacular demonstrations to practical business applications. Despite this focus, companies have not abandoned large models. Apple announced integrating ChatGPT into Siri for more complex tasks, while Microsoft will incorporate OpenAI’s latest model into Windows. However, these integrations are a minor part of their overall AI offerings.

In conclusion, the AI arms race is evolving, with a shift towards smaller, more efficient models that promise to make AI technology more accessible and practical for a wide range of applications.


Secure a 2-year subscription to WSJ and Barron’s News for iOS, Android, PC, and Mac. Enjoy unlimited access to Peggy Noonan’s WSJ insights, Barron’s stock picks, WSJ live TV, audible articles, and Barron’s magazine. Stay informed and subscribe today.

Call Now Button