The engine behind AI factories for the age of AI reasoning—now in full production.
Explore the groundbreaking advancements the NVIDIA Blackwell architecture brings to generative AI and accelerated computing. Building upon generations of NVIDIA technologies, NVIDIA Blackwell defines the next chapter in generative AI with unparalleled performance, efficiency, and scale.
NVIDIA Blackwell-architecture GPUs pack 208 billion transistors and are manufactured using a custom-built TSMC 4NP process. All NVIDIA Blackwell products feature two reticle-limited dies connected by a 10 terabytes per second (TB/s) chip-to-chip interconnect in a unified single GPU.
The second-generation Transformer Engine uses custom NVIDIA Blackwell Tensor Core technology combined with NVIDIA TensorRT?-LLM and NeMo? Framework innovations to accelerate inference and training for large language models (LLMs) and Mixture-of-Experts (MoE) models. NVIDIA Blackwell Tensor Cores add new precisions, including new community-defined microscaling formats, giving high accuracy and ease of replacement for larger precisions.
NVIDIA Blackwell Ultra Tensor Cores are supercharged with 2X the attention-layer acceleration and 1.5X more AI compute FLOPS compared to NVIDIA Blackwell GPUs. The NVIDIA Blackwell Transformer Engine utilizes fine-grain scaling techniques called micro-tensor scaling, to optimize performance and accuracy enabling 4-bit floating point (FP4) AI. This doubles the performance and size of next-generation models that memory can support while maintaining high accuracy.
NVIDIA Blackwell includes NVIDIA Confidential Computing, which protects sensitive data and AI models from unauthorized access with strong hardware-based security. NVIDIA Blackwell is the first TEE-I/O capable GPU in the industry, while providing the most performant confidential compute solution with TEE-I/O capable hosts and inline protection over NVIDIA NVLink?. NVIDIA Blackwell Confidential Computing delivers nearly identical throughput performance compared to unencrypted modes. Enterprises can now secure even the largest models in a performant way, in addition to protecting AI intellectual property (IP) and securely enabling confidential AI training, inference, and federated learning.
Unlocking the full potential of exascale computing and trillion-parameter AI models hinges on the need for swift, seamless communication among every GPU within a server cluster. The fifth-generation of NVIDIA NVLink interconnect can scale up to 576 GPUs to unleash accelerated performance for trillion- and multi-trillion parameter AI models.
The NVIDIA NVLink Switch Chip enables 130TB/s of GPU bandwidth in one 72-GPU NVLink domain (NVL72) and delivers 4X bandwidth efficiency with NVIDIA Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)? FP8 support. The NVIDIA NVLink Switch Chip supports clusters beyond a single server at the same impressive 1.8TB/s interconnect. Multi-server clusters with NVLink scale GPU communications in balance with the increased computing, so NVL72 can support 9X the GPU throughput than a single eight-GPU system.
Data analytics and database workflows have traditionally relied on CPUs for compute. Accelerated data science can dramatically boost the performance of end-to-end analytics, speeding up value generation while reducing cost. Databases, including Apache Spark, play critical roles in handling, processing, and analyzing large volumes of data for data analytics.
NVIDIA Blackwell’s Decompression Engine and ability to access massive amounts of memory in the NVIDIA Grace? CPU over a high-speed link—900 gigabytes per second (GB/s) of bidirectional bandwidth—accelerate the full pipeline of database queries for the highest performance in data analytics and data science with support for the latest compression formats such as LZ4, Snappy, and Deflate.
NVIDIA Blackwell adds intelligent resiliency with a dedicated Reliability, Availability, and Serviceability (RAS) Engine to identify potential faults that may occur early on to minimize downtime. NVIDIA’s AI-powered predictive-management capabilities continuously monitor thousands of data points across hardware and software for overall health to predict and intercept sources of downtime and inefficiency. This builds intelligent resilience that saves time, energy, and computing costs.
NVIDIA’s RAS Engine provides in-depth diagnostic information that can identify areas of concern and plan for maintenance. The RAS engine reduces turnaround time by quickly localizing the source of issues and minimizes downtime by facilitating effective remediation.
Learn more about the architecture that is powering the new era of generative AI and accelerated computing.
什么是行政处罚 | 屁眼火辣辣的疼是什么原因 | 什么异思迁 | 膳食纤维有什么作用 | 为什么瘦不下来 |
阴囊湿疹吃什么药 | 身上出汗多是什么原因 | 神经过敏是什么意思 | 胃反酸吃什么食物好 | 16是什么生肖 |
什么动物没尾巴 | 较前相仿是什么意思 | 黑匣子是什么意思 | 三点水翟读什么 | 水代表什么生肖 |
补充电解质是什么意思 | 上环后同房要注意什么 | 为什么会射精 | 儿童扁桃体发炎吃什么药 | 05年属什么 |
手指麻木是什么病的前兆hcv8jop4ns9r.cn | 梦见自己得绝症了是什么预兆hcv9jop7ns3r.cn | 音色是什么意思hcv8jop8ns3r.cn | 活性炭和木炭有什么区别96micro.com | 九九重阳节是什么意思hcv8jop3ns9r.cn |
梵行是什么意思hcv8jop7ns2r.cn | 色纸是什么hcv9jop2ns9r.cn | 舌系带短挂什么科gysmod.com | 什么是处男hcv9jop3ns8r.cn | 脚底有痣代表什么hcv7jop4ns5r.cn |
什么人会得免疫性脑炎hcv8jop0ns0r.cn | 沉香木是什么树hcv7jop5ns3r.cn | 46是什么意思hcv8jop2ns9r.cn | 甲磺酸倍他司汀片治什么病hcv9jop3ns1r.cn | 为什么牙疼hcv8jop6ns8r.cn |
14岁属什么hcv8jop0ns4r.cn | 睡觉脚抽筋是什么原因引起的ff14chat.com | 什么地流淌hcv8jop8ns4r.cn | 眼袋重是什么原因hcv9jop6ns7r.cn | 6月13日是什么星座hcv7jop6ns7r.cn |