Original Link: https://www.anandtech.com/show/18782/the-nvidia-gtc-spring-2023-keynote-live-blog-800am-pt1500-utc



10:57AM EDT - Welcome to our live blog coverage of NVIDIA’s Spring GTC 2023 keynote address

10:58AM EDT - The traditional kick-off to the show – be it physical or virtual – NVIDIA’s annual spring keynote is showcase for NVIDIA’s vision for the next 12 to 24 months across all of their segments, from graphics to AI to automotive. Along with a slew of product announcements, the presentation, delivered by CEO Jensen Huang always contains a few surprises

10:58AM EDT - Looking at NVIDIA's sizable product stack, NVIDIA is coming off of the launch of their new Hopper and Ada Lovelace GPU architectures for servers and clients respectively. But there are plenty of spots for individual products that remain to be filled. Meanwhile, NVIDIA expects to release their long-awaited Grace CPU this year, and while many of the technical details of that Armv9-based core have since been released, we should hopefully get some launch details for that. As well as its combined CPU+GPU counterpart, Grace Hopper, which places the Grace CPU and Hopper GPU on the same package.

10:58AM EDT -  Meanwhile, we're expecting NVIDIA to take a small victory lap at this year's GTC for having the uncanny timing in launching Hopper and its large language model-friendly Transformer Engines right as the market for GPT and other LLMs has exploded. Now it will be interesting to see how NVIDIA intends to further grow (and profit from) those businesses. The company has all but promissed investors that a cloud service play of some kind will be announced at this GTC.

10:59AM EDT - At this point we're just waiting for the keynote stream to kick off, which should be promptly at 8am

11:00AM EDT - NVIDIA's GTC conference is, bucking the trends, remaining a virtual conference this year

11:01AM EDT - And here we go

11:01AM EDT - Jensen is immediate diving into the subjects of accelerated computing and AI

11:02AM EDT - New advances in NVIDIA's full stack of services

11:02AM EDT - "Welcome to GTC"

11:02AM EDT - Welcoming more than 250K people to this year's virtual conference

11:02AM EDT - 4 years ago the last in-person conference had 8K attendees

11:03AM EDT - "650 amazing talks"

11:04AM EDT - And quickly covering a list of subjects covered in this year's talks

11:05AM EDT - "The purpose of GTC is to inspire the world on the art of what's possible with accelerated computing"

11:05AM EDT - Now rolling NVIDIA's latest "I am AI" video

11:08AM EDT - And, of course, NVIDIA has used AI to put together parts of this video (as they have done for a couple of years now)

11:08AM EDT - "Accelerated computing is not easy"

11:09AM EDT - Accelerated applications can enjoy speed ups and scale ups across many systems

11:09AM EDT - Giving a 1 million-fold increase in performance over the last decade in certain tasks

11:10AM EDT - Using then-and-now comparison of AlexNet and GPT-3 as a comparison of the number of FP operations required to train the respective neural nets

11:10AM EDT - And of course, GPT is all the rage now

11:11AM EDT - "Several thousand applications aare now NVIDIA accelerated"

11:11AM EDT - And NVIDIA has established a cycle of users, applications, and developers to make an active ecosystem

11:12AM EDT - A big part of laying this groundwork has been NVIDIA providing so many libraries for different tasks

11:12AM EDT - Jensen is going through some of those libraries now

11:13AM EDT - NVIDIA is continuing their work with quantum computing and their cuQuantum library

11:13AM EDT - Which is being used to help simulate quantum computers

11:14AM EDT - Today NVIDIA is announcing a quantum control link which allows connecting NV GPUs to quantum computers for error correction (of the quantum computer)

11:15AM EDT - Now on to Spark RAPIDS and vector databases

11:15AM EDT - Introducing a new library: RAFT

11:16AM EDT - For further accelerating vector databases

11:17AM EDT - (As a general reminder, NVIDIA has more software engineers than it does hardware engineers. So new software is a huge part of their total body of work. It also means that software is a huge part of GTC presentations these days)

11:18AM EDT - Jensen is highlighting NVIDIA's partnership with AT&T, which is using NV tech for everything from 5G planning to Riva for voice synthesis

11:18AM EDT - Now talking about NVIDIA's inference platform, recapping TensorRT, Triton, and TMS

11:19AM EDT - New features include multi-GPU, multi-node inference for GPT large language models

11:20AM EDT - Now on to GPU video processing. CV-CUDA and VPF

11:20AM EDT - Microsoft, Tencent, and others are using these libraries to process hundreds of thousands of videos per day

11:21AM EDT - Video processing is a major processing consumer, as a result. Making it a good target for optimization and acceleration

11:21AM EDT - Next up: Genomics

11:21AM EDT - Including NVIDIA Parabricks

11:21AM EDT - Announcing Parabricks 4.1 today

11:22AM EDT - And Holoscan, NV's library for real time medical image processing

11:22AM EDT - And NVIDIA is partnering with Medtronic to develop a common AI platform

11:23AM EDT - Now on to chip manufacturing

11:23AM EDT - Jensen is talking about the extremely small scale of silicon lithography today

11:23AM EDT - Litho is an imaging problem at the edge of physics

11:24AM EDT - Recapping how EUV litho works. And how much it costs - over $250M

11:24AM EDT - As well as how interference patterns within the EUV light is used to create patterns smaller than the light

11:24AM EDT - Computational lithography

11:25AM EDT - Simulating Maxwell's equations to develop and refine litho masks

11:25AM EDT - Computational lithography use is growing fast

11:26AM EDT - New library: cuLitho, to accelerate computational lithography by over 40x

11:26AM EDT - A single reticle currently takes 2 weeks to process. cuLitho can do it in an 8 hour shift

11:26AM EDT - cuLitho can also reduce power consumption by reducing the number of systems required

11:27AM EDT - TSMC will be qualifying cuLitho for production starting in June

11:27AM EDT - Now on to cloud computing

11:28AM EDT - Increasing computing needs are being capped by data center physical power limits, not to mention a desire to cut back on power consumption for environmental reasons

11:28AM EDT - Looks like this will be about NV's Grace CPU

11:28AM EDT - Grace excels where GPUs do not - single threaded serial processing

11:28AM EDT - 72 Arm cores with a 3.2TB/sec fabric

11:29AM EDT - Grace superchip is 2 Graces on a single board

11:29AM EDT - Grace Superchip module

11:29AM EDT - 5 x 8 inches

11:30AM EDT - 2 Grace Superchip modules can fit in a single air-cooled 1U server rack

11:30AM EDT - Claiming 2x the perf at iso-power

11:31AM EDT - Grace is sampling now

11:31AM EDT - And NVIDIA's partners are working to assemble systems

11:31AM EDT - Now on to NVIDIA's networking hardware business

11:31AM EDT - BlueField-3 is in production

11:32AM EDT - That's NVIDIA's latest-generation DPU

11:32AM EDT - Now on to NVIDIA DGX

11:32AM EDT - Half of Fortune 100 companies have installed DGX

11:33AM EDT - Recapping DGX topology and features

11:33AM EDT - DGX H100 is now in full production (now that Intel is finally shipping Sapphire Rapids in volume)

11:34AM EDT - And public cloud providers, including Microsoft's Azure, are quickly adopting DGX for their services

11:34AM EDT - "DGX supercomputers are modern AI factories"

11:34AM EDT - "Generative AI has triggered a sense of urgency to develop AI strategies"

11:34AM EDT - Announcing NVIDIA DGX Cloud

11:35AM EDT - So here's NVIDIA's big cloud services announcement

11:35AM EDT - NVIDIA's ecosystem available via DGX systems hosting cloud instances at the public cloud providers

11:35AM EDT - "Cloud extension of our business model"

11:36AM EDT - Oracle Cloud Infrastructure will be the first DGX public cloud service

11:36AM EDT - 50 early access customers across several industries

11:36AM EDT - Now on to generative AI and its recent explosion

11:37AM EDT - (GPT is going to sell a massive number of H100s at this rate...)

11:38AM EDT - Jensen is recapping large language models and the many things that can be done with GPT and other LLMs, such as generating text and images

11:38AM EDT - "Generative AI is a new kind of computer, one we program in human language"

11:38AM EDT - "Now, everyone is a programmer"

11:39AM EDT - Comparing generative AI to whole platforms such as the PC

11:39AM EDT - Now quickly talking about the many services using generative AI in some form or another

11:39AM EDT - Even accelerated drug design

11:40AM EDT - "The industry needs a foundry. A TSMC for large language models"

11:40AM EDT - Announcing NVIDIA AI Foundations

11:40AM EDT - Language, visual, and biology model making services

11:41AM EDT - Using NVIDIA NeMo, Picasso, and BioNeMo respectively

11:41AM EDT - Customers can create their own models or start with one of NVIDIA's pre-trained models and customize from there

11:42AM EDT - Now rolling a video about AI Foundations and how it works

11:42AM EDT - (This is moving far too quickly to recap it all)

11:43AM EDT - Reinforcement learning used in NeMo to further improve its performance and accuracy

11:43AM EDT - "A personalized, AI model that you control"

11:43AM EDT - That was a video on NeMo. Now we're on to a video about Picasso

11:44AM EDT - Picasso is a service for generating images, video, and models

11:45AM EDT - Getty Images will be using the Picasso service, trained on their library of legally licensed images

11:45AM EDT - Shutterstock will be doing something similar

11:46AM EDT - Announcing a significant expansion of the Adobe partnership to build a set of next-gen AI capabilities into Adobe's software

11:46AM EDT - Adobe Generative Images

11:47AM EDT - And Adobe's Content Authenticity Initiative

11:47AM EDT - Now on to BioNeMo - generative AI for biology

11:48AM EDT - BioNeMo provides models for drug discovery

11:48AM EDT - Protien folding, molecule generation, etc

11:49AM EDT - Accurately predict the structure of a protein in seconds

11:50AM EDT - And that's NVIDIA AI Foundations

11:50AM EDT - Now on to automotive

11:51AM EDT - No, my bad. Now on to talking about data center construction

11:51AM EDT - "No one accelerator can optimally process" the wide variety of models

11:51AM EDT - New inference platform: 4 configurations, 1 architecture, 1 software stack

11:51AM EDT - New product: L4 acceleraetor card. Replaces T4

11:52AM EDT - Want to use L4 to replace CPU servers for AI video processing

11:52AM EDT - Google is offering L4 on Google Cloud

11:53AM EDT - Google GCP is now a premiere NVIDIA AI cloud

11:53AM EDT - More info on that to come later on

11:54AM EDT - L40 accelerator card. This is more of a recap, as the product was released last year

11:54AM EDT - The L series being NVIDIA's designation for server cards based on the Ada Lovelace architecture

11:55AM EDT - L40 is more aimed at image processing/generation, and is the backbone of NVIDIA's Omniverse/OVX hardware

11:55AM EDT - Large language models live up to the name. GPT can get extremely large

11:55AM EDT - Announcing H100 NVL

11:56AM EDT - Dual card/quad slot PCIe product. 2 GH100s with 94GB of memory each

11:56AM EDT - And then Grace Hopper

11:57AM EDT - Grace CPU + Hopper GPU on a single board, for tasks that need both types of processing

11:57AM EDT - CPU/GPU interface 7x faster than PCIe

11:58AM EDT - A bit aspirational at this second, as Grace Hopper is not yet shipping

11:58AM EDT - And that's NVIDIA's AI hardware

11:59AM EDT - Now on to Omniverse

11:59AM EDT - Rolling a video of how Amazon's robotics arm is using Omniverse

12:00PM EDT - Using Isaac Sim to develop the technology

12:01PM EDT - And using simulations to train their models faster

12:02PM EDT - Once again promoting Omniverse and its use of the USD file format

12:02PM EDT - NV has made significant updates to Omniverse in every area

12:02PM EDT - Now rolling a highlight video

12:03PM EDT - DRIVE Sim, Replicator, PhysX Flow, Warp, multi-GPU/multi-node support, Isaac Sim, SimReady Assets, Replicator, Audio2Face, Neural Materials, and more

12:04PM EDT - "Nearly 300K creators and designers have downloaded Omniverse"

12:05PM EDT - Lising numerous new CAD/CAM applications that are now plugged in to Omniverse

12:07PM EDT - And listing off the many companies using Omniverse in some fashion and how they're using it. A lot of manufacturers, to say the least

12:07PM EDT - BMW is building a factory virtually first, 2 years before they build the actual thing

12:09PM EDT - Demoing how a virtual planning session goes

12:09PM EDT - Using Microsoft Teams and Omniverse

12:10PM EDT - Announcing 3 systems to run Omniverse

12:11PM EDT - New generation of workstations powered

12:11PM EDT - New OVX 3.0 servers

12:11PM EDT - And #3: NVIDIA Omniverse Cloud

12:12PM EDT - Omniverse running on cloud services

12:13PM EDT - A fuily-managed cloud service. Partnering with Microsoft, to be hosted on Azure

12:14PM EDT - Connecting Omniverse Cloud to Microsoft's 365 services

12:14PM EDT - Bringing Omniverse to millions of 365 and Azure users

12:15PM EDT - Now recapping the keynote

12:15PM EDT - New hardware, new libraries, and more

12:16PM EDT - Extending business model with NVIDIA DGX Cloud

12:16PM EDT - Best of NVIDIA at best of world's leading CSPs

12:17PM EDT - And NVIDIA AI Foundations for model making services

12:17PM EDT - Plus numerous Omniverse upgrades, and Omniverse cloud services

12:18PM EDT - Jensen is now thanking NVIDIA's partners and employees

12:18PM EDT - And that's a wrap! Please check out our individual articles on NVIDIA's new hardware announcements

12:18PM EDT - https://www.anandtech.com/show/18780/nvidia-announces-h100-nvl-max-memory-server-card-for-large-language-models

12:19PM EDT - https://www.anandtech.com/show/18781/nvidia-unveils-rtx-ada-lovelace-gpus-for-laptops-desktop-rtx-4000-sff

Log in

Don't have an account? Sign up now