AI startup Cerebras debuts 'world's fastest inference' service - with a twist
Cerebras demonstrated how its AI inference can be 10 to 20 times faster than conventional cloud AI inference services. The market for serving up predictions from generative artificial intelligence, what's known as inference, is big business, with OpenAI reportedly on course to collect 3.4 billion in revenue this year serving up predictions by ChatGPT. With a pie that big for inference, there is plenty of room for challengers. On Tuesday, AI chip maker Cerebras Systems of Sunnyvale, California, debuted its AI inference service, which it claims is the fastest in the world and, in many cases, ten to twenty times faster than systems built using the dominant technology, Nvidia's H100 "Hopper" graphics processing unit, or GPU. "We have never seen a technology market growing this fast," said Cebrebras cofounder and CEO Andrew Feldman in a press conference in San Francisco.
Aug-28-2024, 01:04:46 GMT
- Country:
- North America > United States > California
- San Francisco County > San Francisco (0.25)
- Santa Clara County > Sunnyvale (0.24)
- North America > United States > California
- Genre:
- Press Release (0.34)
- Industry:
- Information Technology (1.00)
- Technology: