Wednesday, May 21, 2025
GPTLatest
  • Home
  • GPT News
  • AI News
  • Blog
No Result
View All Result
GPTLatest
  • Home
  • GPT News
  • AI News
  • Blog
No Result
View All Result
GPTLatest
No Result
View All Result
Home AI News

NVIDIA unveils new AI chip to reduce inference costs

August 10, 2023
in AI News
NVIDIA CEO

In a strategic move to maintain its dominance in the fiercely competitive artificial intelligence (AI) hardware sector, NVIDIA Corporation has announced the debut of its groundbreaking AI chip to reduce inference costs. This pivotal step by NVIDIA comes amidst intensifying rivalry from industry giants such as Google, Amazon, and AMD.

The tech giant specialises in graphics processing units, which are currently the preferred chips for LLMs that train generative AI software like OpenAI’s GPT and Google’s Bard. However, Nvidia’s chips are short in supply as startups, tech giants and cloud providers struggle to get GPUs to train their AI models.

The new AI chip, GH 200 Grace Hopper, uses the same GPU capability as Nvidia’s highest-end artificial intelligence chip, H100. However, GH200 combines that GPU capability with 141 GB of advanced memory and a 72-core ARM main processor.

Recap highlights from our special address at #SIGGRAPH2023, including the updated GH200 Grace Hopper Superchip, NVIDIA AI Workbench, and updates on @NVIDIAOmniverse with generative #AI. https://t.co/H925H3ROjo

— NVIDIA (@nvidia) August 8, 2023

GH200 will reduce development costs

CEO Jensen Huang at a conference held on Tuesday said the AI chip would help to scale out of the data centres of the world. He said, “We’re giving this processor a boost.” 

Typically, all AI models require training and inference. The model is initially trained with a large amount of data and requires thousands of General Processing Units. Then the AI model is used to generate content through a process called inference.Inference, unlike training, takes place almost constantly and requires large processing power to run the software.

Huang said, “You can take pretty much any LLM you want and put it in this and it will infer like crazy.” This will ultimately cut down the inference expenses of LLMs.

NVIDIA created GH200 for inference since it has a large memory capacity, providing larger artificial intelligence models to match a single system. The new chip has a 141 GB of memory size, a huge increase from the 80 GB of H100.

“Graphics and artificial intelligence are inseparable, graphics needs AI, and AI needs graphics”

NVIDIA CEO Jensen Huang on its updated GH 200 Grace Hopper Superchip

When will GH200 be available?

Enthusiasts and professionals can anticipate obtaining the AI chip through the company’s authorized distributors during the second quarter of 2024, with sample availability anticipated by the close of 2023. As for the price point, NVIDIA has opted not to disclose these details at this juncture.

This unveiling follows in the wake of a comparable move by NVIDIA’s rival, AMD, which introduced its own AI-focused chip named MI300X, showcasing a substantial memory size of 192 GB. AMD is actively marketing the MI300X for its capacity to effectively support AI inference, setting the stage for a compelling contest between industry giants.

Tags: AIartificiell intelligenschipsnvidia
ShareTweetPin

Related Posts

Tubi
AI News

Tubi Set to Introduce Content Discovery Tool Fueled by OpenAI’s ChatGPT-4

Tubi, the ad-supported TV streaming service owned by Fox, is currently experimenting with a new mobile feature leveraging OpenAI’s GPT-4....

September 30, 2023
Getty Images releases AI-powered photo generator 
AI News

Getty Images releases AI-powered photo generator 

Getty Images has introduced an innovative generative AI art tool. The company, renowned for its extensive collection of editorial photos,...

September 30, 2023
Amazon and Anthropic partner to promote generative AI
AI News

Amazon and Anthropic partner to promote generative AI

Amazon and Anthropic have partnered to broaden the accessibility of Amazon Web Services to customers. The e-commerce giant said it...

September 26, 2023
The Integration of AI in South Korea: From K-pop to Sales Representatives
AI News

The Integration of AI in South Korea: From K-pop to Sales Representatives

AI firm Pulse9 recently introduced Zaein. Zaein is one of South Korea’s most prominent virtual personas to materialize corporate aspirations...

September 26, 2023
Next Post
Bpm

BPM Uses GPT-4 Chatbot to Elevate Team Experiences and Client Value

Azure OpenAI services

Microsoft introduces GPT-35-Turbo and GPT-4 to Azure AI Service 

Recommended

Wolf
GPT News

Hackers Develop XXXGPT and Wolf GPT, New Black Hat AI Tools

August 2, 2023

© 2023 GPTLatest by FiGANT.

  • Home
  • GPT News
  • AI News
  • Blog
This website uses cookies. By continuing to use this website you are giving consent to cookies being used.