A small team of student AI coders beats Google’s machine-learning code

Students from Fast.ai, a small organization that runs free machine-learning courses online, just created an AI algorithm that outperforms code from Google’s researchers, according to an important benchmark.
Fast.ai’s success is important because it sometimes seems as if only those with huge resources can do advanced AI research.
Fast.ai consists of part-time students keen to try their hand at machine learning—and perhaps transition into a career in data science. It rents access to computers in Amazon’s cloud.
But Fast.ai’s team built an algorithm that beats Google’s code, as measured using a benchmark called DAWNBench, from researchers at Stanford. This benchmark uses a common image classification task to track the speed of a deep-learning algorithm per dollar of compute power.
Google’s researchers topped the previous rankings, in a category for training on several machines, using a custom-built collection its own chips designed specifically for machine learning. The Fast.ai team was able to produce something even faster, on roughly equivalent hardware.
“State-of-the-art results are not the exclusive domain of big companies,” says Jeremy Howard, one of Fast.ai’s founders and a prominent AI entrepreneur. Howard and his cofounder, Rachel Thomas, created Fast.ai to make AI more accessible and less exclusive.
Howard’s team was able to compete with the likes of Google by doing a lot of simple things, which are detailed in a blog post. These include making sure that the images fed to its training algorithm were cropped correctly: “These are the obvious, dumb things that many researchers wouldn’t even think to do,” Howard says.
The code needed to run the learning algorithm on several machines was developed by a collaborator at the Pentagon’s new Defense Innovation Unit, created recently to help the military work with AI and machine learning.
Matei Zaharia, a professor at Stanford University and one of the creators of DAWNBench, says the Fast.ai work is impressive, but notes that for many AI tasks large amounts of data and significant compute resources are still key.
The Fast.ai algorithm was trained on the ImageNet database in 18 minutes using 16 Amazon Web Service instances, at a total compute cost of around $40. Howard claims this is about 40 percent better than Google’s effort, although he admits comparison is tricky because the hardware is different.
Jack Clark, director of communications and policy at OpenAI, a nonprofit, says Fast.ai has produced valuable work in other areas such as language understanding. “Things like this benefit everyone because they increase the basic familiarity of people with AI technology,” Clark says.
Deep Dive
Artificial intelligence
Everyone in AI is talking about Manus. We put it to the test.
The new general AI agent from China had some system crashes and server overload—but it’s highly intuitive and shows real promise for the future of AI helpers.
Anthropic can now track the bizarre inner workings of a large language model
What the firm found challenges some basic assumptions about how this technology really works.
China built hundreds of AI data centers to catch the AI boom. Now many stand unused.
The country poured billions into AI infrastructure, but the data center gold rush is unraveling as speculative investments collide with weak demand and DeepSeek shifts AI trends.
OpenAI has released its first research into how using ChatGPT affects people’s emotional well-being
We’re starting to get a better sense of how chatbots are affecting us—but there’s still a lot we don’t know.
Stay connected
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.