• About Us
  • Partnership Opportunities
  • Privacy Policy

Data Center Frontier

Charting the future of data centers and cloud computing.

  • Cloud
    • Hyperscale
  • Colo
    • Site Selection
    • Interconnection
  • Energy
    • Sustainability
  • Cooling
  • Technology
    • Internet of Things
    • AI & Machine Learning
    • Edge Computing
    • Virtual Reality
    • Autonomous Cars
    • 5G Wireless
    • Satellites
  • Design
    • Servers
    • Storage
    • Network
  • Voices
  • Podcast
  • White Papers
  • Resources
    • COVID-19
    • Events
    • Newsletter
    • Companies
    • Data Center 101
  • Jobs
You are here: Home / Machine Learning / The AI Hardware Startups Are Coming. Intel Plans to be Ready

The AI Hardware Startups Are Coming. Intel Plans to be Ready

By Rich Miller - April 29, 2019 Leave a Comment

The AI Hardware Startups Are Coming. Intel Plans to be Ready

Intel's Gadi Singer (left) speaks at the O'Reilly Artificial Intelligence Conference in New York. At right is Sean Gourley, CEO and founder of machine intelligence firm Primer. (Photo: Rich Miller)

LinkedinTwitterFacebookSubscribe
Mail

NEW YORK  – The market for artificial intelligence hardware is entering a new phase, with incumbents competing hard for every workload, even as a cluster of startups prepare to bring specialized AI chips to market.

“At the end of this year, we expect to see a number of companies introduce new hardware for inference with lower energy use,” said Roger Chen, CEO of Computable Labs, said at the kickoff of last week’s O’Reilly Artificial Intelligence Conference, where the new hardware for distributed AI was a hot topic.

“There’s just a stupendous amount of compute coming down the pipe, and it will all be available at the edge,” said Simon Crosby, the Chief Technology Officer at SWIM, which designs software for intelligent edge applications.

The arrival of startup silicon on the AI computing market follows several years of intense competition between chip market leader Intel Corp. and rivals including NVIDIA, AMD and several players advancing ARM technology. Intel continues to hold a dominant position in the enterprise computing space, but the development of powerful new hardware optimized for specific workloads has been a major trend in the high performance computing (HPC) sector, boosted by demand for data-crunching for artificial intelligence and other types of specialized workloads.

Beneficiaries include NVIDIA and AMD, as well as ARM specialist Ampere and a host of startups developing low-power chips to allow smartphones to run AI workloads on the device.

Intel also believes the market is evolving, but expresses confidence in its opportunities in AI, particularly as more compute is applied to inference (decision-making) in addition to training algorithms.

“We are at the start of a new phase,” said Gadi Singer, vice president of the Artificial Intelligence Products Group and general manager of architecture at Intel. “Deep learning, as a technology, is changing the value compute brings to individuals and organizations.

“AI hardware has been advanced,” said Singer. “The focus now goes to inference, and most of the inference in the data center runs on CPUs. This allows companies to use the hardware they have, and add new (accelerator capacity) as they look to enhance the hardware they have.”

The AI Boom Continues

In artificial intelligence (AI), computers are assembled into neural networks that emulate the learning process of the human brain to solve new challenges. It’s a process that requires lots of computing horsepower, which is why the leading players in the field have moved beyond traditional CPU-driven servers.

The race to leverage AI is led by the industry’s marquee names – including Google, Facebook, Amazon and Microsoft – who are seeking to add intelligence to a wide range of services and applications.

Free Resource from Data Center Frontier White Paper Library

cloud service providers
Pro Tips and Best Practices: Physical Layer Strategies for Cloud/Managed Service Providers
Successful Cloud Service Providers and Managed Service Providers need to be out in front of everything in their managed data center spaces – ensuring uptime, bandwidth, and operational/cost efficiency today, with the flexibility and scalability to adapt and expand on the fly. Physical layer and  infrastructure is the foundation on which those services are built. Get the new data center ebook from Siemon that explores pro tips and best practices for physical layer strategies for cloud and managed service providers, from zone cabling in the colocation data center to high speed interconnects in the data center.
We always respect your privacy and we never sell or rent our list to third parties. By downloading this White Paper you are agreeing to our terms of service. You can opt out at any time.

Get this PDF emailed to you.

Venture capitalists and entrepreneurs are also focused on artificial intelligence. Close to 2,000 companies have raised $19 billion in equity to fund startups in AI-related software or services, according to Deepashri Varadharajan, the Lead Analyst for AI at CB Insights. She said one of the most important investment sectors is new hardware.

“Today at least 40 different startups have raised capital to develop AI hardware,” said Varadharajan. “Very narrow focus areas are beginning to emerge, and it is changing the competitive landscape. Five years ago Intel competed with other large manufacturers, like AMD and NVIDIA and ARM.”

The crop of AI hardware startups is spread across the spectrum, including many focusing on edge computing workloads. Here’s an overview of some of the most prominent startups, with links to media coverage:

  • SambaNova Systems: This startup featuring technology from two Stanford researchers Kunle Olukotun and Chris Ré made headlines by raising more than $200 million in its Series A and B rounds, with backing from Intel and Google Ventures.
  • Graphcore: This UK-based AI startup has raised more than $300 million in venture capital and is valued at $1.7 billion. Its investors include BMW and Microsoft, and its technology features a custom chip known as an Intelligence Processing Unit (IPU).
  • Cerebras Systems: This startup, still in stealth mode, is headed by Andrew Feldman, who founded low-energy chip startup SeaMicro and sold it to AMD.  Read more at Barron’s.
  • Habana Labs: This fabless semiconductor startup is developing a purpose-built inference processor. The Next Platform has an overview.
  • Wave Computing: This startup describes its chips as DPUs (DataFlow Processor Units) that can extend an AI platform from the edge to the data center. Here’s a summary from The Next Platform.
  • Mythic: This Austin-based startup making hardware to power AI on devices, and does its processing in flash memory so it can put more power in a device without generating heat and draining the battery.
  • Groq: This stealthy startup was founded by former Google engineers and is reported to be developing a tensor processing unit (TPU), the same class of chip running Google’s AI workloads. It has raised $52 million from Social Capital.
  • FlexLogix: Backed by Lux Capital and Eclipse Ventures, this startup is using an FPGA to target accelerated inference for edge computing. See coverage from ZDNet
  • Cornami: This Santa Clara startup is building chips for both training and inference using a systolic array. See coverage from ZDNet
  • Gyrfalcon Technology: A Milpitas, Calif. startup developing energy-efficient silicon for edge inferencing. See more from ZDNet.
  • Efinex: Another startup focused on edge inferencing for the Internet of Things, leveraging both FPGAs and ASICs. Coverage from ZDNet.

Matching Compute to Specialized Workloads

These AI startups could accelerate the ongoing shift towards specialized computing hardware, featuring custom chips like graphics processing units (GPUs), field programmable gate array (FPGAs) and application-specific integrated circuits (ASICs) in addition to the general purpose x86 CPUs that currently dominate the IT landscape.

This was one of the trends highlighted by researcher Suvojit Ghosh of McMaster University in Hamilton, Ontario, in a recent presentation at DCD New York.

“What we will probably see is a move to silicon that is optimized for workloads,” said Ghosh, who cited GPUs as being particularly cost-efficient in managing dense loads.

Many of these startups aspire to shift computing power and algorithms out of the data center and onto smartphones and tablets, as well as appliances on factory floors and hospitals. The evolution of edge devices and “fog computing” – processing power near the perimeter of the network – will play a role in the geography of the Internet of Things, helping to deliver capacity to billions of devices and sensors.

Some context on chips and AI workloads is helpful in handicapping how this market might evolve beyond traditional CPUs, where Intel has dominated. GPUs have a parallel architecture consisting of hundreds or even thousands of smaller cores designed for handling multiple tasks simultaneously. FPGAs and ASICs are semiconductors that can be customized to perform specialized computing tasks, allowing users to tailor compute power to specific workloads or applications.

There are several types of AI computing workloads. In training, the network learns a new capability from existing data. In inference, the system applies its capabilities to new data, using its training to identify patterns and perform tasks, usually much more quickly than humans could.

Intel’s Plans for a Changing Landscape

Intel’s message is simple: Our platform allows you to get the most mileage from the compute you already operate, while leaving room to hardware accelerators for targeted workloads.

Singer says Intel will introduce several models of its  NNP neural processor accelerator later this year, with the NNP-I 1000 addressing inference, and NNP-L 1000 optimized for training. The NNP I-1000 is currently being used by Facebook and serving “tens of trillions of inference processing transactions,” said Singer.

Singer reinforced the conviction that edge computing is complementary to traditional data centers, and more powerful device-based compute power won’t alter that structure.

“Local data centers are not going to go away,” said Singer. “The growth is going to happen across the whole computing continuum. The scale of inference, along with the economics, argues for having that happen at the edge. There will be more deep learning happening at the edge of the network.

“There is also a strong case for aggregation, which will be seen in data centers,” he added. “We see all of those use cases growing as deep learning becomes a larger component of demand across all these venues. We plan to have products across these compute spaces.”

LinkedinTwitterFacebookSubscribe
Mail

Tagged With: Artificial Intelligence, Intel

Newsletters

Stay informed: Get our weekly updates!

Are you a new reader? Follow Data Center Frontier on Twitter or Facebook.

About Rich Miller

I write about the places where the Internet lives, telling the story of data centers and the people who build them. I founded Data Center Knowledge, the data center industry's leading news site. Now I'm exploring the future of cloud computing at Data Center Frontier.

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Facebook
  • Instagram
  • LinkedIn
  • Pinterest
  • Twitter

Voices of the Industry

5 Ways to Mitigate Supply Chain Unpredictability and Labor Shortages in Data Center Construction

5 Ways to Mitigate Supply Chain Unpredictability and Labor Shortages in Data Center Construction Blake Weaver, Data Center Specialist at ProLift Rigging offers a list of ways to overcome supply chain challenges and labor shortages in data center construction. 

DCF Spotlight

The COVID-19 Crisis and the Data Center Industry

The COVID-19 pandemic presents strategic challenges for the data center and cloud computing sectors. Data Center Frontier provides a one-stop resource for the latest news and analysis for decision-makers navigating this complex new landscape.

An aerial view of major facilities in Data Center Alley in Ashburn, Virginia. (Image: Loudoun County)

Northern Virginia Data Center Market: The Focal Point for Cloud Growth

The Northern Virginia data center market is seeing a surge in supply and an even bigger surge in demand. Data Center Frontier explores trends, stats and future expectations for the No. 1 data center market in the country.

See More Spotlight Features

White Papers

edge

Empower Your Edge Computing — for the Cloud

Today, colocation providers like EdgeConneX are working with leading cloud and IT service providers to bring their solutions to the edge, deployed in close proximity to enterprises, so they can easily access secure, local connectivity to the cloud for any and all of their workloads. Download the new white paper from EdgeConnex exploring edge computing for the cloud that offers an in-depth look at what factors have driven the cloud to the mainstream, the challenges to enterprise cloud adoption, security concerns, spend, the benefits of an integrated cloud solution and more.

Get this PDF emailed to you.

We always respect your privacy and we never sell or rent our list to third parties. By downloading this White Paper you are agreeing to our terms of service. You can opt out at any time.

Newsletters

Get the Latest News from Data Center Frontier

Job Listings

RSS Job Openings | Peter Kazella and Associates, Inc

  • Navy Electrician / Navy Mechanic - Redmond, WA
  • Electrical Commissioning Engineer - Ashburn, VA
  • MEP Superintendent - Data Center - Dallas, TX
  • Construction Project Manager - Data Center - Dallas, TX
  • Data Center QA / QC Manager - Huntsville, AL

See More Jobs

Data Center 101

Data Center 101: Mastering the Basics of the Data Center Industry

Data Center 101: Mastering the Basics of the Data Center Industry

Data Center Frontier, in partnership with Open Spectrum, brings our readers a series that provides an introductory guidebook to the ins and outs of the data center and colocation industry. Think power systems, cooling, solutions, data center contracts and more. The Data Center 101 Special Report series is directed to those new to the industry, or those of our readers who need to brush up on the basics.

  • Data Center Power
  • Data Center Cooling
  • Strategies for Data Center Location
  • Data Center Pricing Negotiating
  • Cloud Computing

See More Data center 101 Topics

About Us

Charting the future of data centers and cloud computing. We write about what’s next for the Internet, and the innovations that will take us there. We tell the story of the digital economy through the data center facilities that power cloud computing and the people who build them. Read more ...
  • Facebook
  • LinkedIn
  • Pinterest
  • Twitter

About Our Founder

Data Center Frontier is edited by Rich Miller, the data center industry’s most experienced journalist. For more than 20 years, Rich has profiled the key role played by data centers in the Internet revolution. Meet the DCF team.

TOPICS

  • 5G Wireless
  • Cloud
  • Colo
  • Connected Cars
  • Cooling
  • Cornerstone
  • Coronavirus
  • Design
  • Edge Computing
  • Energy
  • Executive Roundtable
  • Featured
  • Finance
  • Hyperscale
  • Interconnection
  • Internet of Things
  • Machine Learning
  • Network
  • Podcast
  • Servers
  • Site Selection
  • Social Business
  • Special Reports
  • Storage
  • Sustainability
  • Videos
  • Virtual Reality
  • Voices of the Industry
  • White Paper

Copyright Data Center Frontier LLC © 2021