Nvidia Corp (NASDAQ:NVDA) showcased its H200 graphics processing unit, an upgrade from the H100, the chip OpenAI used to train GPT-4.

Nvidia expects around $16 billion of revenue for its Q3, up 170% year-on-year.

Also Read: What's Going On With Nvidia Monday?

The H100 chips cost between $25,000 - $40,000, CNBC cites from Raymond and James estimates.

The H200 includes 141GB of next-generation "HBM3" memory.

Nvidia said the H200 will generate output nearly twice as fast as the H100 based on a test using Meta Platforms Inc's (NASDAQ: META) Llama 2 LLM.

The H200, which will likely ship in the second quarter of 2024, will compete with Advanced Micro Devices, Inc's (NASDAQ: AMD) MI300X GPU. Amazon.com Inc's (NASDAQ: AMZN) AWS, Alphabet Inc's (NASDAQ: GOOG) (NASDAQ: GOOGL) Google Cloud and Oracle Corp's NYSE: ORCL) Cloud Infrastructure has all committed to using the new chip starting in 2024.

AMD will launch its rival MI300 chip to market in the fourth quarter. Intel Corp (NASDAQ: INTC) claims its Gaudi 2 model is faster than the H100.

Nvidia said the H200 will be compatible with the H100.

The H200 will be available in four-GPU or eight-GPU server configurations on the company's HGX complete systems and in a chip called GH200, which pairs the H200 GPU with an Arm Holdings Plc (NASDAQ: ARM) based processor.

Both the H100 and H200 are based on Nvidia's Hopper architecture.

The stock gained 241% YTD, reaching the $1 trillion valuation thanks to the AI frenzy.

Price Action: NVDA shares traded higher by 0.79% at $487.30 on the last check Monday.

Market News and Data brought to you by Benzinga APIs

Comments
Loading...