IT PARK
    Most Popular

    GPT-4 will allow users to customize the "personality" of the AI, making the avatar a real "person"

    Aug 14, 2023

    Uncover 10 big data myths

    Sep 02, 2023

    How should the edge and the cloud work together?

    Sep 30, 2023

    IT PARK IT PARK

    • Home
    • Encyclopedia

      What is brute force cracking?

      Oct 01, 2023

      What is the reason for the computer card? How to deal with the computer card?

      Sep 30, 2023

      Which is better, laptop, desktop or all-in-one

      Sep 29, 2023

      icloud space is always insufficient to do

      Sep 28, 2023

      What is the difference between the Guid format and MBR format for computer hard drive partitioning?

      Sep 27, 2023
    • AI

      What are the young people interacting with Japan's "Buddhist AI" seeking and escaping from?

      Oct 01, 2023

      Nvidia Announces GH200 Superchip, Most Powerful AI Chip, to Accelerate Generative AI Workloads

      Sep 30, 2023

      Google has categorized 6 real-world AI attacks to prepare for immediately

      Sep 29, 2023

      Samsung considers replacing Google search with Bing AI on devices

      Sep 28, 2023

      Generative AI designs unnatural proteins

      Sep 27, 2023
    • Big Data

      What are the misconceptions in data governance in the digital age?

      Oct 01, 2023

      What is a data warehouse? Why a Data Warehouse?

      Sep 30, 2023

      What is Data Governance? Why do organizations need to do data governance?

      Sep 29, 2023

      Winning Business Excellence with Data Analytics

      Sep 28, 2023

      Has the development of big data come to an end?

      Sep 27, 2023
    • CLO

      How to Reduce the Risk of Cloud Native Applications?

      Oct 01, 2023

      How should the edge and the cloud work together?

      Sep 30, 2023

      Last-generation firewalls won't meet cloud demands

      Sep 29, 2023

      Healthcare Explores Cloud Computing Market: Security Concerns Raise, Multi-Party Collaboration Urgently Needed

      Sep 28, 2023

      Remote work and cloud computing create a variety of endpoint security issues

      Sep 27, 2023
    • IoT

      Berlin showcases smart city innovations

      Oct 01, 2023

      IoT solutions lay the foundation for more effective data-driven policing

      Sep 30, 2023

      CO2 reductions won't happen without digital technology

      Sep 29, 2023

      4 Effective Ways the Internet of Things Can Help with Disaster Management

      Sep 28, 2023

      6 Ways the Internet of Things Can Improve the Lives of Animals

      Sep 27, 2023
    • Blockchain

      Which is better for the logistics industry and blockchain

      Oct 01, 2023

      Will blockchain revolutionize the gaming industry?

      Sep 30, 2023

      How do you make a blockchain investment?

      Sep 29, 2023

      What is the connection between blockchain and Web 3.0?

      Sep 28, 2023

      Canon Launches Ethernet Photo NFT Marketplace Cadabra

      Sep 27, 2023
    IT PARK
    Home » AI » Nvidia Announces GH200 Superchip, Most Powerful AI Chip, to Accelerate Generative AI Workloads
    AI

    Nvidia Announces GH200 Superchip, Most Powerful AI Chip, to Accelerate Generative AI Workloads

    Nvidia announced earlier this Monday that the GH200 Grace Hopper Superchip, Nvidia's most powerful artificial intelligence chip to date, is now in full production.
    Updated: Sep 30, 2023
    Nvidia Announces GH200 Superchip, Most Powerful AI Chip, to Accelerate Generative AI Workloads

    Nvidia's most powerful AI chip to date, the GH200 Grace Hopper Superchip, is now in full production, Nvidia announced earlier this week. the GH200 Superchip is designed to power systems that run the most complex AI workloads, including training the next generation of generative AI models.

    The new chip has a total bandwidth of 900 gigabits per second, seven times more than the standard PCIe Gen5 lanes used in today's most advanced accelerated computing systems. nvidia says the Superchip also consumes five times less power, enabling it to more efficiently handle those demanding AI and high-performance computing applications.

    In particular, the Nvidia GH200 Superchip is expected to be used in generative AI workloads represented by OpenAI ChatGPT, the near-human ability of generative AI to generate new content from prompts that is now sweeping the tech industry.

    Generative AI is rapidly transforming the enterprise, unlocking new opportunities and accelerating discovery in healthcare, finance, business services and many more industries," said Ian Buck, vice president of accelerated computing at Nvidia. With Grace Hopper Superchips in full production, global manufacturers will soon be able to provide enterprises with the acceleration infrastructure they need to build and deploy generative AI applications that employ their unique proprietary data."

    One of the first systems to integrate GH200 Superchips will be Nvidia's own next-generation, large-memory AI supercomputer, the Nvidia DGX GH200. according to Nvidia, this new system uses the NVLink Switch System to combine 256 GH200 Superchips, enabling it to run as a single GPU, delivering up to 1 exaflops of performance (or 1 quintillion floating point operations per second) and 144 TB of shared memory.

    That means it has nearly 500 times more memory and is also more powerful than Nvidia's previous-generation DGX A100 supercomputer, which was launched in 2020 and simply combined eight GPUs into a single chip.

    The DGX GH200 AI supercomputer will also come with a complete full-stack of software for running AI and data analytics workloads, Nvidia said. For example, the system supports Nvidia Base Command software, which provides AI workflow management, cluster management, accelerated compute and storage libraries, as well as network infrastructure and system software. The system also supports Nvidia AI Enterprise, a software layer containing more than 100 AI frameworks, pre-trained models and development tools for streamlining the production of generative AI, computer vision, speech AI and other types of models.

    Constellation Research analyst Holger Mueller said Nvidia has effectively merged two truly reliable products into one by converging Grace and Hopper architectures with NVLink. The result, he said, "is higher performance and capacity, as well as a simplified infrastructure for building AI-driven applications that allows users to see and benefit from so many GPUs and their capabilities as one logical GPU."

    Good things happen when you combine two good things in the right way, and that's the case with Nvidia. the Grace and Hopper chip architectures combined with NVLink not only bring higher performance and capacity, but also simplification for building AI-enabled next-generation applications because of treating all of these GPUs as one logical GPU. "

    The first customers to adopt the new DGX GH200 AI supercomputer include Google Cloud, Meta Platforms and Microsoft, in addition to the DGX GH200 design that Nvidia will make available as a blueprint for cloud service providers who want to customize it for their own infrastructure.

    Girish Bablani, corporate vice president of Azure Infrastructure at Microsoft, said, "Traditionally, training large AI models has been a resource- and time-intensive task, and the potential of the DGX GH200 to handle terabytes of data sets will enable developers to conduct advanced research at a much larger scale and at a much faster pace."

    Nvidia also said it will build the DGX GH200-based AI supercomputer "Nvidia Helios" for its own internal R&D team, which will combine four DGX GH200 systems interconnected using Nvidia Quantum-2 Infiniband networking technology. By the time it goes live at the end of this year, the Helios system will contain a total of 1024 GH200 superchips.

    Finally, Nvidia's server partners are planning to build their own systems based on the new GH200 Superchip, and among the first systems to launch is Quanta Computer's S74G-2U, which will be available later this year.

    Nvidia said server partners have adopted the new Nvidia MGX server specification, which was also announced on Monday. MGX is a modular reference architecture that allows partners to quickly and easily build more than 100 versions of servers based on its latest silicon architecture for a wide range of AI, high-performance computing and other types of workloads. By using NGX, server manufacturers can expect to reduce development costs by as much as three-quarters and cut development time by two-thirds, to about six months.

    NVIDIA Development Chip
    Previous Article Remote work and cloud computing create a variety of endpoint security issues
    Next Article The Future of the Internet of Things and Self-Storage

    Related Articles

    AI

    Developing a new AI project, this is how programming language should be chosen?

    Aug 28, 2023
    Blockchain

    How blockchain technology can be applied to environmental protection to drive a green economy

    Sep 13, 2023
    AI

    AWS releases new product to increase investment in generative AI training

    Aug 24, 2023
    Most Popular

    GPT-4 will allow users to customize the "personality" of the AI, making the avatar a real "person"

    Aug 14, 2023

    Uncover 10 big data myths

    Sep 02, 2023

    How should the edge and the cloud work together?

    Sep 30, 2023
    Copyright © 2023 itheroe.com. All rights reserved. User Agreement | Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.