Sekėjai

Ieškoti šiame dienoraštyje

2026 m. vasario 28 d., šeštadienis

New Nvidia Chip Aims to Speed AI Work --- Amid pressure from rivals, it will offer new product for rapid processing of queries


“Nvidia plans to unveil a new processor specially tailored to help OpenAI and other customers build faster, more efficient tools, a major shake-up to its business that is poised to reset the AI race.

 

The company is designing a new system for so-called "inference" computing, a form of processing that allows AI models to respond to queries, according to people familiar with the plans. The new platform, set to be revealed at Nvidia's GTC developer conference in San Jose, Calif., in March, will incorporate a chip designed by the startup Groq, the people said.

 

Inference computing has been the subject of intense industry competition. Rivals Google and Amazon have designed chips that compete with Nvidia's flagship systems. And the explosion of autonomous coding in the tech workforce has created demand for new chips that can more efficiently handle complex AI-related tasks.

 

OpenAI has agreed to become one of the largest customers of the new processor, some of the people said, representing a major win for Nvidia. The ChatGPT maker, which is one of Nvidia's largest customers, has spent the past few months shopping for more efficient alternatives to Nvidia's chips, and signed a deal with a chip startup in January that provides it with new options.

 

Earlier Friday, OpenAI alluded to the new processor when it announced it would sign up for a major purchase of "dedicated inference capacity" from Nvidia, alongside a $30 billion investment from the chip giant.

 

Nvidia has dominated the business of designing and selling GPUs -- graphics processing units -- a type of processor that can perform billions of simple tasks simultaneously.

 

But for the first time since the start of the AI boom, it is confronting the limits of its flagship product. As the market shifts toward inference, Nvidia is feeling pressure from some customers to produce chips that can more efficiently power AI applications.

 

The chip giant's powerful Hopper, Blackwell and Rubin series GPUs are considered best-in-class for training gigantic AI models and command top prices.

 

Nvidia Chief Executive Jensen Huang has long claimed that Nvidia's GPUs are the market leader for both training and inference, and that such versatility was a key appeal of the product.

 

But over the last year, demand for advanced computing has shifted from training to inference as companies deploy AI agents and other tools that they hope will upend hundreds of industries and generate enormous profits from subscription fees. Agents are AI systems that act relatively autonomously to carry out tasks on behalf of users.” [1]

 

1. U.S. News: New Nvidia Chip Aims to Speed AI Work --- Amid pressure from rivals, it will offer new product for rapid processing of queries. Berber, Jin; Whelan, Robbie; Clark, Kate.  Wall Street Journal, Eastern edition; New York, N.Y.. 28 Feb 2026: A6.  

Komentarų nėra: