How The Nervana Deal Boosts Intel’s Machine Learning Chops

Aug. 11, 2016
Intel’s acquisition of artificial intelligence startup Nervana arms the chipmaker with new software and hardware technologies as it seeks to gain more traction in the machine learning market.

This week Intel expanded its machine learning arsenal with the acquisition of artificial intelligence startup Nervana. The deal provides Intel with new software and hardware technologies as it seeks to gain traction against NVIDIA, whose GPU-based solutions are the platform of choice for many machine learning projects.

The Nervana acquisition also highlights how the growing focus on machine learning is advancing computing technology, which in turn will influence the design and scale of the data centers that house this equipment.

By buying Nervana, Intel adds expertise with ASICs (Application Specific Integrated Circuits) that are highly tailored for machine learning. It’s Intel’s second major acquisition to extend its technology portfolio beyond its core CPU products. Last year Intel acquired Altera, a specialist in field programmable gate arrays (FPGAs), semiconductors that can be reprogrammed to perform specialized tasks. Altera was also a leading player in programmable logic devices (PLDs), which are widely used to automate industrial infrastructure.

Diane Bryant, executive vice president and GM of Intel’s Data Center Group, said the Nervana deal will strengthen Intel’s offerings for high performance computing and hyperscale markets, especially the growing appetite for horsepower for artificial intelligence initiatives in Silicon Valley.

“We believe that by bringing together the Intel engineers who create the Intel Xeon and Intel Xeon Phi processors with the talented Nervana Systems’ team, we will be able to advance the industry faster than would have otherwise been possible,” said Bryant. “We will continue to invest in leading edge technologies that complement and enhance Intel’s AI portfolio.”

Optimizing for Software and Hardware

Nervana CEO Naveen Rao, who founded the company after working on artificial intelligence at Qualcomm, is jazzed about the potential to bring his company’s IT innovations to a larger market.

“We are here to make a dent in the world of computation,” said Rao. “With this deal, we can now shatter the old paradigm and move into a new regime of computing. The combination of Nervana’s technology and expertise incorporated into Intel’s portfolio will take deep learning/AI solutions to the next level. We’ll look back in 10 years and see this time as the inflection point of when compute architectures became neural.

Nervana brings several key capabilities to the table. It uses both software and hardware to optimize data-crunching for artificial intelligence. Its first product is a hosted cloud platform for machine learning, which uses Nervana’s Neon software to accelerate a cluster of NVIDIA TitanX GPUs (graphics processing units). Neon is an open source Python-based language and set of libraries for developing deep learning models.

Nervana is also developing an ASIC that it says has achieved training speeds 10 times faster than conventional GPU-based systems and frameworks. The company says it has built a better interconnect, allowing its Nervana Engine to move data between compute nodes more efficiently. Nervana has been secretive about some of the details of its technology, but The Next Platform has a deeper dive into what’s publicly known.

The Hyperscale Shift

The integration of FPGAs and ASICs is a key strategy for Intel as it sharpens its focus on chips for large data centers. This is driven by the shift to cloud computing, which has shifted buying power from enterprises to large cloud computing platforms.

“Seventy to 80 percent of systems will be deployed in large-scale data centers by 2025,” said Jason Waxman, vice president of the Data Center and Cloud Platforms Group at Intel.

Hyperscale data center operators are seeking to optimize their infrastructures, and FPGAs and ASICs offer more sophisticated processing architectures that can bring HPC-style parallel processing of workloads. This is especially important in machine learning, where neural networks must be trained by absorbing large datasets and refining its algorithms for accuracy. Once a system is trained, it uses inference to “learn” as it encounters new information.

Intel’s Plans for Nervana

Where does Nervana’s technology fit into Intel’s roadmap? Bryant provided the broad outlines of its strategy.

“We will apply Nervana’s software expertise to further optimize the Intel Math Kernel Library and its integration into industry standard frameworks,” said Bryant. “Nervana’s Engine and silicon expertise will advance Intel’s AI portfolio and enhance the deep learning performance and TCO of our Intel Xeon and Intel Xeon Phi processors.”

One benefit might be developing ASIC-powered machine learning platforms for hyperscale customers. Google is a long-time Intel customer that recently developed its own ASIC machine learning hardware (called the Tensor Processing Unit) to power its AI operation. Facebook has also developed its own machine learning servers powered by NVIDIA GPUs.

Nervana’s Rao is confident that with Nervana’s technology, Intel can be a major player in the future of machine learning and AI.

“The semiconductor integrated circuit is one of humanity’s crowning achievements and Intel has the best semiconductor technology in the world,” said Rao. “Nervana’s AI expertise combined with Intel’s capabilities and huge market reach will allow us to realize our vision and create something truly special.”

About the Author

Rich Miller

I write about the places where the Internet lives, telling the story of data centers and the people who build them. I founded Data Center Knowledge, the data center industry's leading news site. Now I'm exploring the future of cloud computing at Data Center Frontier.

Sponsored Recommendations

Guide to Environmental Sustainability Metrics for Data Centers

Unlock the power of Environmental, Social, and Governance (ESG) reporting in the data center industry with our comprehensive guide, proposing 28 key metrics across five categories...

The AI Disruption: Challenges and Guidance for Data Center Design

From large training clusters to small edge inference servers, AI is becoming a larger percentage of data center workloads. Learn more.

A better approach to boost data center capacity – Supply capacity agreements

Explore a transformative approach to data center capacity planning with insights on supply capacity agreements, addressing the impact of COVID-19, the AI race, and the evolving...

How Modernizing Aging Data Center Infrastructure Improves Sustainability

Explore the path to improved sustainability in data centers by modernizing aging infrastructure, uncovering challenges, three effective approaches, and specific examples outlined...

SeventyFour / Shutterstock.com

Improve Data Center Efficiency with Advanced Monitoring and Calculated Points

Max Hamner, Research and Development Engineer at Modius, explains how using calculated points adds up to a superior experience for the DCIM user.

White Papers

Dcf Techno Guard Sr Cover2022 06 15 16 38 10 232x300

High Density IT Cooling

June 21, 2022
Over the last ten years, power density in data centers has been on the rise, leaving many to wonder if the days of air cooling IT equipment are over. This report, courtesy of ...