Search the TechTarget Network
- IT Management
- Industry Sectors
- Technology Topics
KTSDESIGN – STOCK.ADOBE.COM
KTSDESIGN – STOCK.ADOBE.COM
Stanford University finds that AI is outpacing Moore’s Law
Every three months, the speed of artificial intelligence computation doubles, according to Stanford University’s 2019 AI Index report
- Cliff Saran, Managing Editor
Published: 12 Dec 2019 9:56
Stanford University’s AI Index 2019 annual report has found that the speed of artificial intelligence (AI) is outpacing Moore’s Law.DOWNLOAD THIS FREE GUIDE
How AI is moulding the future of unified communications
Download this e-guide and take a look at the rise of AI within unified communications, the main benefits of UCaaS for enterprises and the ever-evolving world of enterprise collaboration.
- Corporate E-mail Address:
- I agree to my information being processed by TechTarget and its Partners to contact me via phone, email, or other means regarding information relevant to my professional interests. I may unsubscribe at any time.
Moore’s Law maps out how processor speeds double every 18 months to two years, which means application developers can expect a doubling in application performance for the same hardware cost.
But the Stanford report, produced in partnership with McKinsey & Company, Google, PwC, OpenAI, Genpact and AI21Labs, found that AI computational power is accelerating faster than traditional processor development. “Prior to 2012, AI results closely tracked Moore’s Law, with compute doubling every two years.,” the report said. “Post-2012, compute has been doubling every 3.4 months.”
The study looked at how AI algorithms have improved over time, by tracking the progress of the ImageNet image identification program. Given that image classification methods are largely based on supervised machine learning techniques, the report’s authors looked at how long it takes to train an AI model and associated costs, which they said represents a measurement of the maturity of AI development infrastructure, reflecting advances in software and hardware.
Their research found that over 18 months, the time required to train a network on cloud infrastructure for supervised image recognition fell from about three hours in October 2017 to about 88 seconds in July 2019. The report noted that data on ImageNet training time on private cloud instances was in line with the public cloud AI training time improvements.
The report’s authors used the ResNet image classification model to assess how long it takes algorithms to achieve a high level of accuracy. In October 2017, 13 days of training time were required to reach just above 93% accuracy. The report found that training an AI-based image classification over 13 days to achieve 93% accuracy would have cost about $2,323 in 2017.
The study reported that the latest benchmark available on Stanford DAWNBench , using a cloud TPU on GCP to run the ResNet model to attain image classification accuracy slightly above 93% accuracy, cost just over $12 in September 2018.
Read more about artificial intelligence
- Google Cloud has expanded its committed use discount plan to include GPU, TPU and local SSD resources to spark more AI and machine learning workloads.
- Huawei aims to speed up AI training times with the launch of a new processor in China and a new AI computing framework.
The report also explored how far computer vision had progressed, looking at innovative algorithms that push the limits of automatic activity understanding, which can recognise human actions and activities from videos using the ActivityNet Challenge.
One of the tasks in this challenge, called Temporal Activity Localisation, uses a long video sequences that depict more than one activity, and the algorithm is asked to find a given activity. Today, algorithms can accurately recognise hundreds of complex human activities in real time, but the report found that much more work is needed.
“After organising the International Activity Recognition Challenge (ActivityNet) for the last four years, we observe that more research is needed to develop methods that can reliably discriminate activities, which involve fine-grained motions and/or subtle patterns in motion cues, objects and human-object interactions,” said Bernard Ghanem, associate professor of electrical engineering at King Abdullah University of Science and Technology, in the report.
“Looking forward, we foresee the next generation of algorithms to be one that accentuates learning without the need for excessively large manually curated data. In this scenario, benchmarks and competitions will remain a cornerstone to track progress in this self-learning domain.”
Read more on Artificial intelligence, automation and robotics
How to achieve explainability in AI modelsFuture Decoded: AI-powered data revolutionises software developmentHow to optimize storage for AI, machine learning and deep learning3 important steps to get started with AIAI for decision-making shows promise, but worker trust an issuecomputer hallucinationLatest News
- Alarm bells ring, the IoT is listening
- New government faces calls to urgently deliver on pre-election pledge to review IR35 reforms
- Subpostmasters got the best deal possible in legal battle with the Post Office, says lawyer
- View All News
Download Computer Weekly
- IN THE CURRENT ISSUE:
- General Election 2019: Which manifesto is best for the UK’s tech sector?
- Alternative databases set for mainstream adoption?
- Overcoming the container security conundrum: What enterprises need to know
Latest Blog Posts
- Not 4 sale– Downtime
- Vilified then vindicated – victory for subpostmasters in Post Office trial shows risk of tech hubris– Computer Weekly Editors Blog
- View All Blogs
- An inside look at Alibaba’s deep learning processor– ComputerWeekly.com
- How Amadeus used FPGAs to break Moore’s Law threshold– ComputerWeekly.com
- How to optimize storage for AI, machine learning and …– SearchStorage
Start the conversation
Send me notifications when other members comment.Add My Comment
- Transforming operations for successful cloud adoptionStill considering making the move to the cloud? Here are best practices and cloud-centric processes CIOs should follow to enable …
- IT workforce skews younger, says analysis of US dataThe IT workforce is getting younger, according to government IT workforce data. The reasons for this are subject to a debate that…
- Top edge computing trends to watch in 2020Edge computing is still an evolving technology domain and enterprises should expect continued evolution in the year ahead. Here …
- About Us
- Meet The Editors
- Contact Us
- Business Partners
- Media Kit
- Corporate Site
All Rights Reserved, Copyright 2000 – 2019, TechTarget