Close
  • Latest News
  • Artificial Intelligence
  • Video
  • Big Data and Analytics
  • Cloud
  • Networking
  • Cybersecurity
  • Applications
  • IT Management
  • Storage
  • Sponsored
  • Mobile
  • Small Business
  • Development
  • Database
  • Servers
  • Android
  • Apple
  • Innovation
  • Blogs
  • PC Hardware
  • Reviews
  • Search Engines
  • Virtualization
Read Down
Sign in
Close
Welcome!Log into your account
Forgot your password?
Read Down
Password recovery
Recover your password
Close
Search
Logo
Subscribe
Logo
  • Latest News
  • Artificial Intelligence
  • Video
  • Big Data and Analytics
  • Cloud
  • Networking
  • Cybersecurity
  • Applications
  • IT Management
  • Storage
  • Sponsored
  • Mobile
  • Small Business
  • Development
  • Database
  • Servers
  • Android
  • Apple
  • Innovation
  • Blogs
  • PC Hardware
  • Reviews
  • Search Engines
  • Virtualization
More
    Subscribe
    Home Latest News

      AI Scaling Laws Face Diminishing Returns, Pushing Labs Toward New AI Model Training Strategies

      Written by

      Madeline Clarke
      Published December 10, 2024
      Share
      Facebook
      Twitter
      Linkedin
        Digital illustration of machine learning and artificial intelligence concept.
        Image: Lee/Adobe Stock

        eWEEK content and product recommendations are editorially independent. We may make money when you click on links to our partners. Learn More.

        Developers have begun seeking new AI training methods as recent models have started to see shrinking returns. According to recent reports, models created within top AI labs are experiencing slower improvement rates than they used to, prompting developers to question the validity of AI scaling laws for model training. Improving model training means rethinking the guiding strategies surrounding our understanding of artificial intelligence model enhancement.

        For the past five years, developers operated under the belief that pretraining models using more compute and data would produce more powerful AI capabilities. However, recent reports point to slower improvement rates for AI models, even from top AI labs. Google’s AI model Gemini failed to achieve performance gains, and Anthropic experienced development issues that prompted a delay in the release of its Claude 3.5 Opus AI model. Similarly, OpenAI’s Orion model isn’t meeting performance expectations regarding coding improvements and saw minimal new enhancements compared to previous models.

        With models showing declining enhancement rates and diminished returns, AI developers could be forced to enter a new age of model training methods.

        The Problem With AI Scaling Laws

        Recent AI model developments show that AI labs can’t rely solely on applying more data and computation to produce more enhanced models. This poses a challenge for labs that have used traditional AI scaling laws as a leading factor in their model development operations. While the initial versions of AI models produced by developers like OpenAI, Meta, Anthropic, and Google had improved by combining more GPUs and larger data quantities, these methods alone cannot sustain exponential growth.

        AI scaling laws had also previously contributed to expectations for the technology’s seemingly endless enhancement potential. Top AI companies have made significant investments in model development and made optimistic claims regarding AI’s future that reflect the anticipated returns granted from these technologies. In response, developers are applying new training strategies to remain competitive as we enter the next era of AI scaling laws.

        New Training Strategies To Pick Up The Pace

        AI labs are working quickly to improve models through new scaling methods. Recently favored strategies include “test-time compute,” a training technique that gives models more time to compute to produce output to considered questions. A recent paper published by MIT researchers shows that test-time compute substantially improves model performance on reasoning tasks.

        Developers will likely continue to use traditional methods, such as applying more relevant datasets and compute clusters. However, the demand for fast AI interference chips could increase if the time-test compute strategy becomes the next leading training method. Whatever the future of AI development may be, the overall buzz around its potential remains positive as AI labs compete to produce the next best thing in the AI model scene.

        Madeline Clarke
        Madeline Clarke
        Madeline is a writer specializing in copywriting and content creation. After studying Art and earning her BFA in Creative Writing at Salisbury University she applied her knowledge of writing and design to develop creative and influential copy. She has since formed her business, Clarke Content, LLC, through which she produces entertaining, informational content and represents companies with professionalism and taste.

        Get the Free Newsletter!

        Subscribe to Daily Tech Insider for top news, trends & analysis

        Get the Free Newsletter!

        Subscribe to Daily Tech Insider for top news, trends & analysis

        MOST POPULAR ARTICLES

        Artificial Intelligence

        9 Best AI 3D Generators You Need...

        Sam Rinko - June 25, 2024 0
        AI 3D Generators are powerful tools for many different industries. Discover the best AI 3D Generators, and learn which is best for your specific use case.
        Read more
        Cloud

        RingCentral Expands Its Collaboration Platform

        Zeus Kerravala - November 22, 2023 0
        RingCentral adds AI-enabled contact center and hybrid event products to its suite of collaboration services.
        Read more
        Artificial Intelligence

        8 Best AI Data Analytics Software &...

        Aminu Abdullahi - January 18, 2024 0
        Learn the top AI data analytics software to use. Compare AI data analytics solutions & features to make the best choice for your business.
        Read more
        Latest News

        Zeus Kerravala on Networking: Multicloud, 5G, and...

        James Maguire - December 16, 2022 0
        I spoke with Zeus Kerravala, industry analyst at ZK Research, about the rapid changes in enterprise networking, as tech advances and digital transformation prompt...
        Read more
        Video

        Datadog President Amit Agarwal on Trends in...

        James Maguire - November 11, 2022 0
        I spoke with Amit Agarwal, President of Datadog, about infrastructure observability, from current trends to key challenges to the future of this rapidly growing...
        Read more
        Logo

        eWeek has the latest technology news and analysis, buying guides, and product reviews for IT professionals and technology buyers. The site’s focus is on innovative solutions and covering in-depth technical content. eWeek stays on the cutting edge of technology news and IT trends through interviews and expert analysis. Gain insight from top innovators and thought leaders in the fields of IT, business, enterprise software, startups, and more.

        Facebook
        Linkedin
        RSS
        Twitter
        Youtube

        Advertisers

        Advertise with TechnologyAdvice on eWeek and our other IT-focused platforms.

        Advertise with Us

        Menu

        • About eWeek
        • Subscribe to our Newsletter
        • Latest News

        Our Brands

        • Privacy Policy
        • Terms
        • About
        • Contact
        • Advertise
        • Sitemap
        • California – Do Not Sell My Information

        Property of TechnologyAdvice.
        © 2024 TechnologyAdvice. All Rights Reserved

        Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.

        ×