AI news for: Hardware
Explore AI news and udpates focusing on Hardware for the last 7 days.

Alibaba to offer Nvidia’s physical AI development tools in its AI platform
Alibaba said on Wednesday that it is integrating Nvidia's AI development tools for robotics, self-driving cars and connected spaces into its Cloud Pla...

Just Released: NVIDIA HPC SDK v25.9
The new release introduces support for CUDA 13.0, and includes updated library components, bug fixes, and performance improvements....

Key Takeaways:
- CUDA version 13.0 is bundled, alongside CUDA 12.9 in some packages.
- Compatible with various Linux distributions, including RHEL/Rocky, SLES/SUSE, and Ubuntu.
- Installation instructions are provided for various Linux platforms and package managers.

Nvidia is letting anyone use its AI voice animation tech
Nvidia is open-sourcing Audio2Face, its AI-powered tool that generates realistic facial animations for 3D avatars — all based on audio input. The chan...

Key Takeaways:
- Developers can now use the Audio2Face tool and its underlying framework to create realistic 3D characters for their games and apps.
- The tool allows developers to create 3D characters for both pre-scripted content and livestreams.
- The training frame of Audio2Face is being made available, enabling users to tweak its models for different use cases.

Open Secret: How NVIDIA Nemotron Models, Datasets and Techniques Fuel AI Development
Open technologies — made available to developers and businesses to adopt, modify and innovate with — have been part of every major technology shift, f...

Key Takeaways:
- Nemotron provides a transparent and adaptable foundation for building AI applications, enabling developers to understand how their models work and trust the results.
- The technology supports both generalized intelligence and specialized intelligence, making it suitable for a wide range of AI use cases across industries.
- Nemotron has been adopted by various companies, including CrowdStrike, DataRobot, and ServiceNow, for their AI development needs, and its dataset is also used to inform the design of NVIDIA's future AI systems.

Deploy High-Performance AI Models in Windows Applications on NVIDIA RTX AI PCs
Today, Microsoft is making Windows ML available to developers. Windows ML enables C#, C++ and Python developers to optimally run AI models locally acr...

Key Takeaways:
- Windows ML unlocks full TensorRT acceleration for GeForce RTX and RTX Pro GPUs, delivering exceptional AI performance on Windows 11.
- TensorRT for RTX Execution Provider provides 50% faster throughput compared to prior DirectML implementations on NVIDIA RTX GPUs and supports low-latency inference.
- Windows ML developers can leverage architecture advancements like FP8 and FP4 on the Tensor Cores and support a variety of model architectures, including LLMs, diffusion, CNN, and more with the new ONNX Runtime APIs.

Build a Retrieval-Augmented Generation (RAG) Agent with NVIDIA Nemotron
Unlike traditional LLM-based systems that are limited by their training data, retrieval-augmented generation (RAG) improves text generation by incorpo...

Key Takeaways:
- Agentic RAG allows LLMs to access external knowledge, adapt to changing information, and perform complex reasoning tasks dynamically.
- The new approach can be built using NVIDIA NIM endpoints and tools like LangGraph, enabling developers to create advanced AI agents.
- Agentic RAG has applications in various domains, including customer support, IT Help Desks, and more, where agents can retrieve and process relevant information to provide accurate responses.

Nvidia Invests in OpenAI With $100 Billion to Build Out More AI Data Centers - CNET
Nvidia Invests in OpenAI With $100 Billion to Build Out More AI Data Centers CNETNvidia stock jumps on $100 billion OpenAI investment as Huang touts '...

Key Takeaways:
- OpenAI will deploy at least 10 gigawatts of Nvidia systems for AI data centers over the coming years.
- Nvidia's investment will grow in scale as each new system is deployed, with the first phase scheduled for 2026.
- This partnership positions Nvidia at the center of the AI boom, buying into the biggest AI company, and could define how quickly AI advances in the future.

China’s ‘full stack’ AI ambitions on show at Alibaba’s Apsara Conference - South China Morning Post
China’s ‘full stack’ AI ambitions on show at Alibaba’s Apsara Conference South China Morning PostAlibaba Stock’s AI-Powered Run Isn’t Done Barron'sAli...

R²D²: Three Neural Breakthroughs Transforming Robot Learning from NVIDIA Research
While today's robots excel in controlled settings, they still struggle with the unpredictability, dexterity, and nuanced interactions required for rea...

Key Takeaways:
- NeRD (Neural Robot Dynamics) enables accurate dynamics prediction, with less than 0.1% error in accumulated reward for 1,000-step policy evaluation.
- Reference-Scoped Exploration (RSE) streamlines learning dexterous manipulation from human demonstrations, achieving almost 20% more success rates with the Inspire hand.
- VT-Refine combines vision and touch for robust bimanual assembly, improving real-world success rates by approximately 20% in the vision-only variant and 40% for the visuo-tactile variant.

Faster Training Throughput in FP8 Precision with NVIDIA NeMo
In previous posts on FP8 training, we explored the fundamentals of FP8 precision and took a deep dive into the various scaling recipes for practical l...

Key Takeaways:
- FP8 training yields a speedup of up to 1.53x for large language models (Llama 3.1 405B) compared to BF16 on NVIDIA H100 GPUs.
- MXFP8 recipe on DGX B200 GPUs achieves a consistent speedup of 1.28x to 1.37x across different model sizes, with a peak speedup of 1.37x.
- Larger models experience more pronounced efficiency gains with FP8, as the reduced memory footprint and higher throughput benefit more significantly from reduced-precision arithmetic.

How to Accelerate Community Detection in Python Using GPU-Powered Leiden
Community detection algorithms play an important role in understanding data by identifying hidden groups of related entities in networks. Social netwo...

Key Takeaways:
- cuGraph's GPU-accelerated Leiden implementation outperforms comparable CPU implementations by up to 47x.
- Leiden from cuGraph is easily accessible in Python workflows through the cuGraph library or NetworkX library through the nx-cugraph backend.
- GPU-powered Leiden workflows can scale large data and solve bigger problems in far less time, making it ideal for genomics and other applications relying on community detection.

Build a Real-Time Visual Inspection Pipeline with NVIDIA TAO 6 and NVIDIA DeepStream 8
Building a robust visual inspection pipeline for defect detection and quality control is not easy. Manufacturers and developers often face challenges ...

Key Takeaways:
- NVIDIA TAO enables the customization of vision foundation models using domain adaptation, self-supervised fine-tuning, and knowledge distillation.
- DeepStream 8 provides a low-code tool for deploying model ideas into standalone applications or microservices with improved inference accuracy and throughput.
- Together, TAO and DeepStream 8 accelerate model development, improve accuracy, and enable real-time visual inspection pipelines for various industries such as automotive, logistics, and manufacturing.

Nvidia and Abu Dhabi institute launch joint AI and robotics lab in the UAE - Reuters
Nvidia and Abu Dhabi institute launch joint AI and robotics lab in the UAE ReutersNvidia, Abu Dhabi Institute Open AI and Robotics Lab in UAE Yahoo Fi...

Key Takeaways:
- The joint lab will explore AI applications in areas such as climate, energy, and genomics.
- The TII will use Nvidia's Thor chip to advance robotics research and development.
- Specific use cases are planned for robotic innovations, including transportation and logistics.

How to GPU-Accelerate Model Training with CUDA-X Data Science
In previous posts on AI in manufacturing and operations, we covered the unique data challenges in the supply chain and how smart feature engineering c...

Key Takeaways:
- Tree-based models are ideal for structured manufacturing data, outperforming neural networks in this domain.
- Popular library XGBoost can be 150x faster in inference using the Forest Inference Library from NVIDIA cuML, making it suitable for production use.
- Feature importance analysis helps engineers understand variable relationships and identify noise features, offering a robust way to filter out uninformative data.

Canada Goes All In on AI: NVIDIA Joins Nations’ Technology Leaders in Montreal to Shape Sovereign AI Strategy
Canada’s role as a leader in artificial intelligence was on full display at this week’s All In Canada AI Ecosystem event. NVIDIA Vice President of Gen...

Key Takeaways:
- Canada's AI factory, built by TELUS and powered by NVIDIA, offers end-to-end AI capabilities and ensures full data residency and control within Canadian borders.
- The country's AI momentum is accelerated through collaborations between public and private sector leaders, shaping national infrastructure, innovation, and policy for AI development.
- RBC Capital Markets is using NVIDIA software to build enterprise-grade AI agents for capital markets research, enabling Canadian institutions to deploy intelligent systems tailored to local needs.

At Climate Week NYC, NVIDIA Details AI’s Key Role in the Sustainable Energy Transition
Energy efficiency in large language model inference has improved 100,000x in the past 10 years — demonstrating that accelerated computing is sustainab...

Key Takeaways:
- Forecasted AI-induced energy savings across energy-intensive sectors by 2035: 4.5% of projected energy demand will be saved.
- Emerald AI and NVIDIA collaborate on a reference design for grid-friendly, energy-efficient AI infrastructure, unlocking 100 gigawatts of untapped power grid capacity.
- NVIDIA's product carbon footprint summary comparison reveals a 24% reduction in embodied carbon emissions intensity between NVIDIA HGX H100 and HGX B200 baseboards.