<?xml version="1.0" encoding="UTF-8"?><feed
  xmlns="http://www.w3.org/2005/Atom"
  xmlns:thr="http://purl.org/syndication/thread/1.0"
  xml:lang=""
  xml:base="https://developer.nvidia.com/blog/wp-atom.php"
   >
	<title type="text">NVIDIA Technical Blog</title>
	<subtitle type="text">News and tutorials for developers, data scientists, and IT admins</subtitle>

	<updated>2026-05-14T19:27:31Z</updated>

	<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog" />
	<id>https://developer.nvidia.com/blog/feed/</id>
	<link rel="self" type="application/atom+xml" href="https://developer.nvidia.com/blog/feed/" />

	
		<entry>
		<author>
			<name>Graham Steele</name>
					</author>
		<title type="html"><![CDATA[How the NVIDIA Vera Rubin Platform is Solving Agentic AI’s Scale-Up Problem]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-the-nvidia-vera-rubin-platform-is-solving-agentic-ais-scale-up-problem/" />
		<id>https://developer.nvidia.com/blog/?p=116892</id>
		<updated>2026-05-14T19:27:31Z</updated>
		<published>2026-05-14T19:24:35Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Groq 3 LPX" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" fetchpriority="high" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="(max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" />Agentic inference has fundamentally changed the runtime dynamics of inference workloads by introducing non-deterministic trajectories—actions, observations,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-the-nvidia-vera-rubin-platform-is-solving-agentic-ais-scale-up-problem/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="(max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" />Agentic inference has fundamentally changed the runtime dynamics of inference workloads by introducing non-deterministic trajectories—actions, observations,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" /><p>Agentic inference has fundamentally changed the runtime dynamics of inference workloads by introducing non-deterministic trajectories—actions, observations, and decisions that an AI agent produces while working through a task. These trajectories compound end-to-end latency across hundreds of inference requests per session. NVIDIA Vera Rubin NVL72 handles the bulk of that inference load as…</p>
<p><a href="https://developer.nvidia.com/blog/how-the-nvidia-vera-rubin-platform-is-solving-agentic-ais-scale-up-problem/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-the-nvidia-vera-rubin-platform-is-solving-agentic-ais-scale-up-problem/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-the-nvidia-vera-rubin-platform-is-solving-agentic-ais-scale-up-problem/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Samuel Ochoa</name>
					</author>
		<title type="html"><![CDATA[Transform Video Into Instantly Searchable, Actionable Intelligence with AI Agents and Skills ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/transform-video-into-instantly-searchable-actionable-intelligence-with-ai-agents-and-skills/" />
		<id>https://developer.nvidia.com/blog/?p=116772</id>
		<updated>2026-05-14T19:09:52Z</updated>
		<published>2026-05-13T18:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Computer Vision / Video Analytics" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="Blueprint" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Metropolis" /><category scheme="https://developer.nvidia.com/blog" term="News" /><category scheme="https://developer.nvidia.com/blog" term="VLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-social-nurec-devpage-kv-li-1600x900" />In today’s data-driven world, organizations increasingly rely on video to capture critical information, yet extracting meaningful, real-time insights from...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/transform-video-into-instantly-searchable-actionable-intelligence-with-ai-agents-and-skills/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-social-nurec-devpage-kv-li-1600x900" />In today’s data-driven world, organizations increasingly rely on video to capture critical information, yet extracting meaningful, real-time insights from...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/robotics-social-nurec-devpage-kv-li-1600x900-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-social-nurec-devpage-kv-li-1600x900" /><p>In today’s data-driven world, organizations increasingly rely on video to capture critical information, yet extracting meaningful, real-time insights from massive amounts of footage remains a challenge. NVIDIA Metropolis Blueprint for video search and summarization (VSS) overcomes this hurdle by transforming millions of live video streams or hours of recorded video into instantly searchable…</p>
<p><a href="https://developer.nvidia.com/blog/transform-video-into-instantly-searchable-actionable-intelligence-with-ai-agents-and-skills/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/transform-video-into-instantly-searchable-actionable-intelligence-with-ai-agents-and-skills/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/transform-video-into-instantly-searchable-actionable-intelligence-with-ai-agents-and-skills/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Irina Demeshko</name>
					</author>
		<title type="html"><![CDATA[Accelerated X-Ray Analysis for Nanoscale Imaging (XANI) of Novel Materials]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/accelerated-x-ray-analysis-for-nanoscale-imaging-xani-of-novel-materials/" />
		<id>https://developer.nvidia.com/blog/?p=116726</id>
		<updated>2026-05-14T18:52:16Z</updated>
		<published>2026-05-13T16:39:20Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GPUDirect" /><category scheme="https://developer.nvidia.com/blog" term="NVL72" /><category scheme="https://developer.nvidia.com/blog" term="Python" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="hpc-accelerated-x-ray-analysis" />A massive-scale X-ray free-electron laser (XFEL) enables tracking structural and electron dynamics in novel systems, including fusion materials, semiconductors,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/accelerated-x-ray-analysis-for-nanoscale-imaging-xani-of-novel-materials/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="hpc-accelerated-x-ray-analysis" />A massive-scale X-ray free-electron laser (XFEL) enables tracking structural and electron dynamics in novel systems, including fusion materials, semiconductors,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/hpc-accelerated-x-ray-analysis.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="hpc-accelerated-x-ray-analysis" /><p>A massive-scale X-ray free-electron laser (XFEL) enables tracking structural and electron dynamics in novel systems, including fusion materials, semiconductors, batteries, and catalysis. It produces ultrashort X-ray pulses that can record the movements of atoms and electrons. These instruments can detect the smallest change in material structure caused by defects and other influences.</p>
<p><a href="https://developer.nvidia.com/blog/accelerated-x-ray-analysis-for-nanoscale-imaging-xani-of-novel-materials/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/accelerated-x-ray-analysis-for-nanoscale-imaging-xani-of-novel-materials/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/accelerated-x-ray-analysis-for-nanoscale-imaging-xani-of-novel-materials/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Lovina Dmello</name>
					</author>
		<title type="html"><![CDATA[How to Eliminate Pipeline Friction in AI Model Serving]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-eliminate-pipeline-friction-in-ai-model-serving/" />
		<id>https://developer.nvidia.com/blog/?p=116526</id>
		<updated>2026-05-14T18:52:18Z</updated>
		<published>2026-05-12T18:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="Dynamo-Triton" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="ONNX" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="tensorrt-optimized-industries" />The path from a trained AI model to production should be smooth, but rarely is. Many teams invest weeks fine-tuning models, only to discover that exporting to a...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-eliminate-pipeline-friction-in-ai-model-serving/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="tensorrt-optimized-industries" />The path from a trained AI model to production should be smooth, but rarely is. Many teams invest weeks fine-tuning models, only to discover that exporting to a...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/tensorrt-optimized-industries-1.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="tensorrt-optimized-industries" /><p>The path from a trained AI model to production should be smooth, but rarely is. Many teams invest weeks fine-tuning models, only to discover that exporting to a deployment format breaks layers, input shapes cause runtime failures, or version mismatches silently degrade performance. These issues are collectively known as pipeline friction, and they cost organizations time, money…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-eliminate-pipeline-friction-in-ai-model-serving/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-eliminate-pipeline-friction-in-ai-model-serving/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-eliminate-pipeline-friction-in-ai-model-serving/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Christian Shrauder</name>
					</author>
		<title type="html"><![CDATA[Introducing NVIDIA Fleet Intelligence for Real-Time GPU Fleet Visibility and Optimization]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/introducing-nvidia-fleet-intelligence-for-real-time-gpu-fleet-visibility-and-optimization/" />
		<id>https://developer.nvidia.com/blog/?p=116707</id>
		<updated>2026-05-14T18:52:19Z</updated>
		<published>2026-05-11T19:44:27Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Confidential Compute" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center (2)" />The compute capability of large GPU fleets presents unprecedented opportunities to innovate and provide value to customers in record time. Yet these...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/introducing-nvidia-fleet-intelligence-for-real-time-gpu-fleet-visibility-and-optimization/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center (2)" />The compute capability of large GPU fleets presents unprecedented opportunities to innovate and provide value to customers in record time. Yet these...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/data-center-2.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center (2)" /><p>The compute capability of large GPU fleets presents unprecedented opportunities to innovate and provide value to customers in record time. Yet these advancements come with a variety of challenges. At scale, teams are juggling heterogeneous hardware, fast‑moving software stacks, tight power envelopes, and spiky, multitenant workloads. A single hotspot, misconfigured driver, or subtle hardware fault…</p>
<p><a href="https://developer.nvidia.com/blog/introducing-nvidia-fleet-intelligence-for-real-time-gpu-fleet-visibility-and-optimization/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/introducing-nvidia-fleet-intelligence-for-real-time-gpu-fleet-visibility-and-optimization/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/introducing-nvidia-fleet-intelligence-for-real-time-gpu-fleet-visibility-and-optimization/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Joseph Lucas</name>
					</author>
		<title type="html"><![CDATA[Improving Bash Generation in Small Language Models with Grammar-Constrained Decoding]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/improving-bash-generation-in-small-language-models-with-grammar-constrained-decoding/" />
		<id>https://developer.nvidia.com/blog/?p=116680</id>
		<updated>2026-05-14T18:52:20Z</updated>
		<published>2026-05-08T17:13:33Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI / Cybersecurity" /><category scheme="https://developer.nvidia.com/blog" term="AI Red Team" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Security for AI" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-960x540-jpg.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-jpg.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="person-desk-three-computers" />Bash is one of the most flexible and powerful interfaces exposed to AI agents. In the right system, a model that emits grep, curl, tar, or a shell pipeline is...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/improving-bash-generation-in-small-language-models-with-grammar-constrained-decoding/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-960x540-jpg.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-jpg.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="person-desk-three-computers" />Bash is one of the most flexible and powerful interfaces exposed to AI agents. In the right system, a model that emits grep, curl, tar, or a shell pipeline is...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-960x540-jpg.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/person-desk-three-computers-jpg.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="person-desk-three-computers" /><p>Bash is one of the most flexible and powerful interfaces exposed to AI agents. In the right system, a model that emits , , , or a shell pipeline is producing an executable action that can read files, mutate a workspace, open network connections, and chain tools together. For the NVIDIA AI Red Team, this makes command generation a useful research target. If smaller language models can be guided…</p>
<p><a href="https://developer.nvidia.com/blog/improving-bash-generation-in-small-language-models-with-grammar-constrained-decoding/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/improving-bash-generation-in-small-language-models-with-grammar-constrained-decoding/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/improving-bash-generation-in-small-language-models-with-grammar-constrained-decoding/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Matej Kosec</name>
					</author>
		<title type="html"><![CDATA[Streaming Tokens and Tools: Multi-Turn Agentic Harness Support in NVIDIA Dynamo ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/streaming-tokens-and-tools-multi-turn-agentic-harness-support-in-nvidia-dynamo/" />
		<id>https://developer.nvidia.com/blog/?p=116658</id>
		<updated>2026-05-14T18:52:21Z</updated>
		<published>2026-05-08T15:59:16Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MoE nvidia technical blog" />An agentic exchange must preserve a structured interaction: assistant turns interleave reasoning with one or more tool calls, and subsequent user turns return...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/streaming-tokens-and-tools-multi-turn-agentic-harness-support-in-nvidia-dynamo/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MoE nvidia technical blog" />An agentic exchange must preserve a structured interaction: assistant turns interleave reasoning with one or more tool calls, and subsequent user turns return...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/MoE-nvidia-technical-blog-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MoE nvidia technical blog" /><p>An agentic exchange must preserve a structured interaction: assistant turns interleave reasoning with one or more tool calls, and subsequent user turns return the corresponding tool results to the model context. Reasoning replay is model- and turn-dependent: some reasoning should be retained, while some should be dropped. The inference engine is responsible for supporting this more expressive…</p>
<p><a href="https://developer.nvidia.com/blog/streaming-tokens-and-tools-multi-turn-agentic-harness-support-in-nvidia-dynamo/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/streaming-tokens-and-tools-multi-turn-agentic-harness-support-in-nvidia-dynamo/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/streaming-tokens-and-tools-multi-turn-agentic-harness-support-in-nvidia-dynamo/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Felix Abecassis</name>
					</author>
		<title type="html"><![CDATA[Achieving Peak System and Workload Efficiency on NVIDIA GB200 NVL72 with Slurm Block Scheduling]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/achieving-peak-system-and-workload-efficiency-on-nvidia-gb200-nvl72-with-slurm-block-scheduling/" />
		<id>https://developer.nvidia.com/blog/?p=116606</id>
		<updated>2026-05-14T18:52:22Z</updated>
		<published>2026-05-07T21:20:14Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Slurm" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-gb200-nvl72" />NVIDIA GB200 NVL72 introduces a fundamentally new way to build GPU clusters by extending NVIDIA NVLink coherence across an entire rack. This design enables...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/achieving-peak-system-and-workload-efficiency-on-nvidia-gb200-nvl72-with-slurm-block-scheduling/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-gb200-nvl72" />NVIDIA GB200 NVL72 introduces a fundamentally new way to build GPU clusters by extending NVIDIA NVLink coherence across an entire rack. This design enables...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/nvidia-gb200-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-gb200-nvl72" /><p>NVIDIA GB200 NVL72 introduces a fundamentally new way to build GPU clusters by extending NVIDIA NVLink coherence across an entire rack. This design enables exascale performance, but it also changes the assumptions that many scheduling systems were built on. As a result, “rack-scale locality” becomes a hard constraint. When workloads cross domain boundaries, performance drops sharply…</p>
<p><a href="https://developer.nvidia.com/blog/achieving-peak-system-and-workload-efficiency-on-nvidia-gb200-nvl72-with-slurm-block-scheduling/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/achieving-peak-system-and-workload-efficiency-on-nvidia-gb200-nvl72-with-slurm-block-scheduling/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/achieving-peak-system-and-workload-efficiency-on-nvidia-gb200-nvl72-with-slurm-block-scheduling/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ruixiang Wang</name>
					</author>
		<title type="html"><![CDATA[Model Quantization: Post-Training Quantization Using NVIDIA Model Optimizer]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/model-quantization-post-training-quantization-using-nvidia-model-optimizer/" />
		<id>https://developer.nvidia.com/blog/?p=116649</id>
		<updated>2026-05-14T18:52:24Z</updated>
		<published>2026-05-07T21:18:06Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="Model Optimizer" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube-column" />Model quantization is an effective method to reduce VRAM usage and improve inference performance on consumer devices such as NVIDIA GeForce RTX GPUs. By...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/model-quantization-post-training-quantization-using-nvidia-model-optimizer/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube-column" />Model quantization is an effective method to reduce VRAM usage and improve inference performance on consumer devices such as NVIDIA GeForce RTX GPUs. By...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/cube-column.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube-column" /><p>Model quantization is an effective method to reduce VRAM usage and improve inference performance on consumer devices such as NVIDIA GeForce RTX GPUs. By lowering computational and memory requirements while preserving model quality, quantization helps AI models run more efficiently in resource-constrained environments. This post walks through how to use NVIDIA Model Optimizer to quantize a…</p>
<p><a href="https://developer.nvidia.com/blog/model-quantization-post-training-quantization-using-nvidia-model-optimizer/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/model-quantization-post-training-quantization-using-nvidia-model-optimizer/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/model-quantization-post-training-quantization-using-nvidia-model-optimizer/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ava Arnaz</name>
					</author>
		<title type="html"><![CDATA[Real-Time Performance Monitoring and Faster Debugging with NCCL Inspector and Prometheus]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/real-time-performance-monitoring-and-faster-debugging-with-nccl-inspector-and-prometheus/" />
		<id>https://developer.nvidia.com/blog/?p=116529</id>
		<updated>2026-05-14T18:52:25Z</updated>
		<published>2026-05-07T16:02:58Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Accelerated Computing Libraries" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring.webp 1024w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring" />Distributed deep learning depends on fast, reliable GPU-to-GPU communication using the NVIDIA Collective Communication Library (NCCL). When training slows down,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/real-time-performance-monitoring-and-faster-debugging-with-nccl-inspector-and-prometheus/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring.webp 1024w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring" />Distributed deep learning depends on fast, reliable GPU-to-GPU communication using the NVIDIA Collective Communication Library (NCCL). When training slows down,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring.webp 1024w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVIDIA-NCCL-Inspector-Real-Time-Performance-Monitoring" /><p>Distributed deep learning depends on fast, reliable GPU-to-GPU communication using the NVIDIA Collective Communication Library (NCCL). When training slows down, it becomes challenging to determine why and what to do next. A problem can span computation, communication, a specific rank, or underlying hardware. NVIDIA NCCL Inspector accelerates triaging by providing a lightweight and continuous…</p>
<p><a href="https://developer.nvidia.com/blog/real-time-performance-monitoring-and-faster-debugging-with-nccl-inspector-and-prometheus/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/real-time-performance-monitoring-and-faster-debugging-with-nccl-inspector-and-prometheus/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/real-time-performance-monitoring-and-faster-debugging-with-nccl-inspector-and-prometheus/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Felix Friedmann</name>
					</author>
		<title type="html"><![CDATA[How to Build In-Vehicle AI Agents with NVIDIA: From Cloud to Car ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-in-vehicle-ai-agents-with-nvidia-from-cloud-to-car/" />
		<id>https://developer.nvidia.com/blog/?p=116326</id>
		<updated>2026-05-14T18:52:26Z</updated>
		<published>2026-05-05T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="Automotive / Transportation" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="Robotics Compute" /><category scheme="https://developer.nvidia.com/blog" term="TensorRT-LLM" /><category scheme="https://developer.nvidia.com/blog" term="Thor" /><category scheme="https://developer.nvidia.com/blog" term="VLMs" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="FeatureImage" />The automotive cockpit is undergoing a fundamental shift from rule-based interfaces to agentic, multimodal AI systems capable of reasoning, planning, and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-build-in-vehicle-ai-agents-with-nvidia-from-cloud-to-car/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="FeatureImage" />The automotive cockpit is undergoing a fundamental shift from rule-based interfaces to agentic, multimodal AI systems capable of reasoning, planning, and...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/FeatureImage.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="FeatureImage" /><p>The automotive cockpit is undergoing a fundamental shift from rule-based interfaces to agentic, multimodal AI systems capable of reasoning, planning, and acting. In most vehicles on the road today, in-vehicle assistants still rely on fixed command-response patterns: interpret a phrase, trigger an action, reset. While effective for well-defined tasks, this approach doesn’t scale to modern…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-build-in-vehicle-ai-agents-with-nvidia-from-cloud-to-car/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-in-vehicle-ai-agents-with-nvidia-from-cloud-to-car/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-build-in-vehicle-ai-agents-with-nvidia-from-cloud-to-car/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Eduardo Alvarez</name>
					</author>
		<title type="html"><![CDATA[Building for the Rising Complexity of Agentic Systems with Extreme Co-Design]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-for-the-rising-complexity-of-agentic-systems-with-extreme-co-design/" />
		<id>https://developer.nvidia.com/blog/?p=116408</id>
		<updated>2026-05-14T19:15:37Z</updated>
		<published>2026-05-05T15:52:15Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Groq 3 LPX" /><category scheme="https://developer.nvidia.com/blog" term="NemoClaw" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />Generative AI’s explosive first chapter was defined by humans sending requests and models responding. The agentic chapter is different.&nbsp; Agents don't...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-for-the-rising-complexity-of-agentic-systems-with-extreme-co-design/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />Generative AI’s explosive first chapter was defined by humans sending requests and models responding. The agentic chapter is different.&nbsp; Agents don't...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/05/image3.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" /><p>Generative AI’s explosive first chapter was defined by humans sending requests and models responding. The agentic chapter is different. Agents don’t follow a pre-determined sequence of actions. They call tools, spawn sub-agents with different tasks and models, retain information in memory, manage their own context window, and decide for themselves when they’re finished. In doing so…</p>
<p><a href="https://developer.nvidia.com/blog/building-for-the-rising-complexity-of-agentic-systems-with-extreme-co-design/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-for-the-rising-complexity-of-agentic-systems-with-extreme-co-design/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-for-the-rising-complexity-of-agentic-systems-with-extreme-co-design/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Adi Geva</name>
					</author>
		<title type="html"><![CDATA[Optimize Supply Chain Decision Systems Using NVIDIA cuOpt Agent Skills]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/optimize-supply-chain-decision-systems-using-nvidia-cuopt-agent-skills/" />
		<id>https://developer.nvidia.com/blog/?p=115751</id>
		<updated>2026-05-14T18:52:29Z</updated>
		<published>2026-05-04T20:55:05Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Agentic-AI" />Modern supply chains operate under the constant pressures of fluctuating demand, volatile costs, constrained capacity, and interdependent decision-making....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/optimize-supply-chain-decision-systems-using-nvidia-cuopt-agent-skills/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Agentic-AI" />Modern supply chains operate under the constant pressures of fluctuating demand, volatile costs, constrained capacity, and interdependent decision-making....<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Agentic-AI.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Agentic-AI" /><p>Modern supply chains operate under the constant pressures of fluctuating demand, volatile costs, constrained capacity, and interdependent decision-making. Traditionally, specialized operations research (OR) teams solved these problems by translating business questions into mathematical models. This process can take weeks and often produces fragile solutions that struggle to adapt when conditions…</p>
<p><a href="https://developer.nvidia.com/blog/optimize-supply-chain-decision-systems-using-nvidia-cuopt-agent-skills/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/optimize-supply-chain-decision-systems-using-nvidia-cuopt-agent-skills/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/optimize-supply-chain-decision-systems-using-nvidia-cuopt-agent-skills/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Homam Bahnassi</name>
					</author>
		<title type="html"><![CDATA[Speed Up Unreal Engine NNE Inference with NVIDIA TensorRT for RTX Runtime]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/speed-up-unreal-engine-nne-inference-with-nvidia-tensorrt-for-rtx-runtime/" />
		<id>https://developer.nvidia.com/blog/?p=116176</id>
		<updated>2026-05-14T18:52:32Z</updated>
		<published>2026-04-30T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Unreal Engine" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="TensorRT-RTX" />Neural network techniques are increasingly used in computer graphics to boost image quality, improve performance, and streamline content creation. Approaches...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/speed-up-unreal-engine-nne-inference-with-nvidia-tensorrt-for-rtx-runtime/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="TensorRT-RTX" />Neural network techniques are increasingly used in computer graphics to boost image quality, improve performance, and streamline content creation. Approaches...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/05/TensorRT-RTX.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="TensorRT-RTX" /><p>Neural network techniques are increasingly used in computer graphics to boost image quality, improve performance, and streamline content creation. Approaches like super resolution, denoising, and neural rendering help real-time engines work more efficiently, offering new creative possibilities while keeping performance in mind. Unreal Engine 5 (UE5) has taken several steps in this direction…</p>
<p><a href="https://developer.nvidia.com/blog/speed-up-unreal-engine-nne-inference-with-nvidia-tensorrt-for-rtx-runtime/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/speed-up-unreal-engine-nne-inference-with-nvidia-tensorrt-for-rtx-runtime/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/speed-up-unreal-engine-nne-inference-with-nvidia-tensorrt-for-rtx-runtime/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Phillip Singh</name>
					</author>
		<title type="html"><![CDATA[Build AI-Powered Games with NVIDIA DLSS 4.5, RTX, and Unreal Engine 5]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-ai-powered-games-with-nvidia-dlss-4-5-rtx-and-unreal-engine-5/" />
		<id>https://developer.nvidia.com/blog/?p=116210</id>
		<updated>2026-05-14T18:52:30Z</updated>
		<published>2026-04-30T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="ACE" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="DLSS" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Gaming" /><category scheme="https://developer.nvidia.com/blog" term="GeForce" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Neural Graphics" /><category scheme="https://developer.nvidia.com/blog" term="NvRTX" /><category scheme="https://developer.nvidia.com/blog" term="Ray Tracing / Path Tracing" /><category scheme="https://developer.nvidia.com/blog" term="RTX Kit" /><category scheme="https://developer.nvidia.com/blog" term="Text Processing" /><category scheme="https://developer.nvidia.com/blog" term="Unreal Engine" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image4" />Today, game developers can begin integrating NVIDIA DLSS 4.5 with Dynamic Multi Frame Generation, Multi Frame Generation 6X, and the second-generation...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-ai-powered-games-with-nvidia-dlss-4-5-rtx-and-unreal-engine-5/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image4" />Today, game developers can begin integrating NVIDIA DLSS 4.5 with Dynamic Multi Frame Generation, Multi Frame Generation 6X, and the second-generation...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image4-6.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image4" /><p>Today, game developers can begin integrating NVIDIA DLSS 4.5 with Dynamic Multi Frame Generation, Multi Frame Generation 6X, and the second-generation transformer model for NVIDIA Super Resolution. In this post, we’ll go over new technologies and resources to share with our game-developer community, including: At CES 2026, we introduced DLSS 4.5, extending its AI-driven…</p>
<p><a href="https://developer.nvidia.com/blog/build-ai-powered-games-with-nvidia-dlss-4-5-rtx-and-unreal-engine-5/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-ai-powered-games-with-nvidia-dlss-4-5-rtx-and-unreal-engine-5/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-ai-powered-games-with-nvidia-dlss-4-5-rtx-and-unreal-engine-5/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Joel Pennington</name>
					</author>
		<title type="html"><![CDATA[How to Build, Run, and Scale High-Quality Creator Workflows in ComfyUI]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-run-and-scale-high-quality-creator-workflows-in-comfyui/" />
		<id>https://developer.nvidia.com/blog/?p=116284</id>
		<updated>2026-05-14T18:52:32Z</updated>
		<published>2026-04-30T16:16:04Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="AI workflows" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GeForce" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="RTX GPU" /><category scheme="https://developer.nvidia.com/blog" term="Tutorial" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="workstation-tech-blog-600x338" />Creative and visualization teams today produce more assets, in more formats, with leaner teams. Generative AI can accelerate that work – compressing tasks...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-build-run-and-scale-high-quality-creator-workflows-in-comfyui/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="workstation-tech-blog-600x338" />Creative and visualization teams today produce more assets, in more formats, with leaner teams. Generative AI can accelerate that work – compressing tasks...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/workstation-tech-blog-600x338-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="workstation-tech-blog-600x338" /><p>Creative and visualization teams today produce more assets, in more formats, with leaner teams. Generative AI can accelerate that work – compressing tasks that once took hours of manual effort into automated, repeatable pipelines. ComfyUI is an open-source, node-based creative tool that runs locally on NVIDIA RTX GPUs. It connects image generation, video synthesis, and language models into…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-build-run-and-scale-high-quality-creator-workflows-in-comfyui/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-run-and-scale-high-quality-creator-workflows-in-comfyui/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-build-run-and-scale-high-quality-creator-workflows-in-comfyui/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Zhengyi Zhang</name>
					</author>
		<title type="html"><![CDATA[Automating GPU Kernel Translation with AI Agents: cuTile Python to cuTile.jl]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/automating-gpu-kernel-translation-with-ai-agents-cutile-python-to-cutile-jl/" />
		<id>https://developer.nvidia.com/blog/?p=116188</id>
		<updated>2026-05-14T18:52:34Z</updated>
		<published>2026-04-30T15:54:38Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="CUDA Tile" /><category scheme="https://developer.nvidia.com/blog" term="cuTile" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on code on their computer." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Automate-Agent-Workflow" />NVIDIA CUDA Tile (cuTile) is a tile-based programming model that enables developers to write GPU kernels in terms of tile-level operations—loads, stores, and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/automating-gpu-kernel-translation-with-ai-agents-cutile-python-to-cutile-jl/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on code on their computer." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Automate-Agent-Workflow" />NVIDIA CUDA Tile (cuTile) is a tile-based programming model that enables developers to write GPU kernels in terms of tile-level operations—loads, stores, and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on code on their computer." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Automate-Agent-Workflow.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Automate-Agent-Workflow" /><p>NVIDIA CUDA Tile (cuTile) is a tile-based programming model that enables developers to write GPU kernels in terms of tile-level operations—loads, stores, and matrix multiply-accumulate—rather than manually coordinating threads, warps, and shared memory. cuTile.jl brings the same tile-based approach to the dynamic programming language Julia. Users can write custom GPU kernels without dropping…</p>
<p><a href="https://developer.nvidia.com/blog/automating-gpu-kernel-translation-with-ai-agents-cutile-python-to-cutile-jl/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/automating-gpu-kernel-translation-with-ai-agents-cutile-python-to-cutile-jl/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/automating-gpu-kernel-translation-with-ai-agents-cutile-python-to-cutile-jl/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Shashank Sabhlok</name>
					</author>
		<title type="html"><![CDATA[Powering AI Factories with NVIDIA Enterprise Reference Architectures]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/powering-ai-factories-with-nvidia-enterprise-reference-architectures/" />
		<id>https://developer.nvidia.com/blog/?p=116222</id>
		<updated>2026-05-14T18:52:35Z</updated>
		<published>2026-04-29T16:41:57Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB300" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on a data center rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="AI-Factory" />The next wave of enterprise productivity is being built on AI factories. As organizations deploy agentic AI systems capable of reasoning, automation, and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/powering-ai-factories-with-nvidia-enterprise-reference-architectures/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on a data center rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="AI-Factory" />The next wave of enterprise productivity is being built on AI factories. As organizations deploy agentic AI systems capable of reasoning, automation, and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="A person working on a data center rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/AI-Factory.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="AI-Factory" /><p>The next wave of enterprise productivity is being built on AI factories. As organizations deploy agentic AI systems capable of reasoning, automation, and real-time decision-making at scale, competitive advantage increasingly depends on the infrastructure that supports them. Success requires more than raw compute. It demands a scalable, predictable foundation that can orchestrate intelligent…</p>
<p><a href="https://developer.nvidia.com/blog/powering-ai-factories-with-nvidia-enterprise-reference-architectures/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/powering-ai-factories-with-nvidia-enterprise-reference-architectures/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/powering-ai-factories-with-nvidia-enterprise-reference-architectures/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Dejun Lin</name>
					</author>
		<title type="html"><![CDATA[Scaling Biomolecular Modeling Using Context Parallelism in NVIDIA BioNeMo]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/scaling-biomolecular-modeling-using-context-parallelism-in-nvidia-bionemo/" />
		<id>https://developer.nvidia.com/blog/?p=115250</id>
		<updated>2026-05-14T18:52:36Z</updated>
		<published>2026-04-28T19:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Drug Discovery" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="HPC / Scientific Computing" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="biomolecule" />For decades, computational biology has operated under a reductionist compromise. To fit complex biological systems into the limited memory of a single GPU,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/scaling-biomolecular-modeling-using-context-parallelism-in-nvidia-bionemo/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="biomolecule" />For decades, computational biology has operated under a reductionist compromise. To fit complex biological systems into the limited memory of a single GPU,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/biomolecule-1.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="biomolecule" /><p>For decades, computational biology has operated under a reductionist compromise. To fit complex biological systems into the limited memory of a single GPU, researchers have had to deconstruct them into isolated fragments—single proteins or small domains. This created a context gap, where larger proteins or complexes could not be folded zero-shot due to GPU hardware memory constraints. Now…</p>
<p><a href="https://developer.nvidia.com/blog/scaling-biomolecular-modeling-using-context-parallelism-in-nvidia-bionemo/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/scaling-biomolecular-modeling-using-context-parallelism-in-nvidia-bionemo/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/scaling-biomolecular-modeling-using-context-parallelism-in-nvidia-bionemo/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anjali Shah</name>
					</author>
		<title type="html"><![CDATA[NVIDIA Nemotron 3 Nano Omni Powers Multimodal Agent Reasoning in a Single Efficient Open Model]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model/" />
		<id>https://developer.nvidia.com/blog/?p=116077</id>
		<updated>2026-05-14T18:52:37Z</updated>
		<published>2026-04-28T16:01:12Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decoratove image showing mult-modal processing." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Nemotron-3-Nano-Omni" />Agentic systems often reason across screens, documents, audio, video, and text within a single perception‑to‑action loop. However, they still rely on...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decoratove image showing mult-modal processing." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Nemotron-3-Nano-Omni" />Agentic systems often reason across screens, documents, audio, video, and text within a single perception‑to‑action loop. However, they still rely on...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decoratove image showing mult-modal processing." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Nemotron-3-Nano-Omni.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Nemotron-3-Nano-Omni" /><p>Agentic systems often reason across screens, documents, audio, video, and text within a single perception‑to‑action loop. However, they still rely on fragmented model chains—separate stacks for vision, audio, and text. This increases inference hops and orchestration complexity, driving up inference costs while weakening cross-modal context consistency. NVIDIA Nemotron 3 Nano Omni…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Tsubasa Onishi</name>
					</author>
		<title type="html"><![CDATA[24/7 Simulation Loops: How Agentic AI Keeps Subsurface Engineering Moving]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/24-7-simulation-loops-how-agentic-ai-keeps-subsurface-engineering-moving/" />
		<id>https://developer.nvidia.com/blog/?p=116079</id>
		<updated>2026-05-14T18:52:39Z</updated>
		<published>2026-04-28T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="Energy" /><category scheme="https://developer.nvidia.com/blog" term="Energy Exploration &amp; Generation" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LangChain" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" />The subsurface industry is at a critical point in its digital evolution. For decades, unlocking reservoir potential has relied on experts performing essential...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/24-7-simulation-loops-how-agentic-ai-keeps-subsurface-engineering-moving/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" />The subsurface industry is at a critical point in its digital evolution. For decades, unlocking reservoir potential has relied on experts performing essential...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image2-8.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" /><p>The subsurface industry is at a critical point in its digital evolution. For decades, unlocking reservoir potential has relied on experts performing essential and time-intensive manual workflows. As data complexity grows, the gap between machine speed and human bandwidth has become a primary bottleneck. On-demand simulation workflows are currently hampered by both manual data overhead…</p>
<p><a href="https://developer.nvidia.com/blog/24-7-simulation-loops-how-agentic-ai-keeps-subsurface-engineering-moving/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/24-7-simulation-loops-how-agentic-ai-keeps-subsurface-engineering-moving/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/24-7-simulation-loops-how-agentic-ai-keeps-subsurface-engineering-moving/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anu Srivastava</name>
					</author>
		<title type="html"><![CDATA[Build with DeepSeek V4 Using NVIDIA Blackwell and GPU-Accelerated Endpoints]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/" />
		<id>https://developer.nvidia.com/blog/?p=116127</id>
		<updated>2026-05-14T18:52:40Z</updated>
		<published>2026-04-24T23:29:56Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-model-representation-2" />DeepSeek just launched its fourth generation of flagship models with DeepSeek-V4-Pro and DeepSeek-V4-Flash, both targeted at enabling highly efficient...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-model-representation-2" />DeepSeek just launched its fourth generation of flagship models with DeepSeek-V4-Pro and DeepSeek-V4-Flash, both targeted at enabling highly efficient...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ai-model-representation-2.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-model-representation-2" /><p>DeepSeek just launched its fourth generation of flagship models with DeepSeek-V4-Pro and DeepSeek-V4-Flash, both targeted at enabling highly efficient million-token context inference. DeepSeek-V4-Pro is the largest model in the family, with 1.6T total parameters and 49B active parameters. DeepSeek-V4-Flash is a smaller 284B-parameter model with 13B active parameters, designed for higher-speed…</p>
<p><a href="https://developer.nvidia.com/blog/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
<link href="https://developer.download.nvidia.com/video/devblog/Limerick.mov" rel="enclosure" length="10" type="video/quicktime" />
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Holger Roth</name>
					</author>
		<title type="html"><![CDATA[Federated Learning Without the Refactoring Overhead Using NVIDIA FLARE]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/" />
		<id>https://developer.nvidia.com/blog/?p=116007</id>
		<updated>2026-05-14T18:52:41Z</updated>
		<published>2026-04-24T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Federated Learning" /><category scheme="https://developer.nvidia.com/blog" term="Internet/Communications" /><category scheme="https://developer.nvidia.com/blog" term="NVFLARE" /><category scheme="https://developer.nvidia.com/blog" term="NVIDIA Flare" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Connected healthcare facilities graphic" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-500x282.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-195x110.jpg 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="connected-healthcare-facilities-graphic" />Federated learning (FL) is no longer a research curiosity—it’s a practical response to a hard constraint: the most valuable data is often the least movable....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Connected healthcare facilities graphic" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-500x282.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-195x110.jpg 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="connected-healthcare-facilities-graphic" />Federated learning (FL) is no longer a research curiosity—it’s a practical response to a hard constraint: the most valuable data is often the least movable....<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Connected healthcare facilities graphic" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic.jpg 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-500x282.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2023/06/connected-healthcare-facilities-graphic-195x110.jpg 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="connected-healthcare-facilities-graphic" /><p>Federated learning (FL) is no longer a research curiosity—it’s a practical response to a hard constraint: the most valuable data is often the least movable. Regulatory boundaries, data sovereignty rules, and organizational risk tolerance routinely prevent centralized aggregation. Meanwhile, sheer data gravity makes even permitted transfers slow, expensive, and fragile at scale.</p>
<p><a href="https://developer.nvidia.com/blog/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Chris Deotte</name>
						<uri>https://www.kaggle.com/cdeotte</uri>
					</author>
		<title type="html"><![CDATA[Winning a Kaggle Competition with Generative AI–Assisted Coding]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/winning-a-kaggle-competition-with-generative-ai-assisted-coding/" />
		<id>https://developer.nvidia.com/blog/?p=116054</id>
		<updated>2026-05-14T18:52:43Z</updated>
		<published>2026-04-23T20:15:02Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Kaggle" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" />In March 2026, three LLM agents generated over 600,000 lines of code, ran 850 experiments, and helped secure a first-place finish in a Kaggle playground...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/winning-a-kaggle-competition-with-generative-ai-assisted-coding/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" />In March 2026, three LLM agents generated over 600,000 lines of code, ran 850 experiments, and helped secure a first-place finish in a Kaggle playground...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" /><p>In March 2026, three LLM agents generated over 600,000 lines of code, ran 850 experiments, and helped secure a first-place finish in a Kaggle playground competition. Success in modern machine learning competitions is increasingly defined by how quickly you can generate, test, and iterate on ideas. LLM agents, combined with GPU acceleration, dramatically compress this loop. Historically…</p>
<p><a href="https://developer.nvidia.com/blog/winning-a-kaggle-competition-with-generative-ai-assisted-coding/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/winning-a-kaggle-competition-with-generative-ai-assisted-coding/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/winning-a-kaggle-competition-with-generative-ai-assisted-coding/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Aart J.C. Bik</name>
					</author>
		<title type="html"><![CDATA[Simplify Sparse Deep Learning with Universal Sparse Tensor in nvmath-python]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/simplify-sparse-deep-learning-with-universal-sparse-tensor-in-nvmath-python/" />
		<id>https://developer.nvidia.com/blog/?p=114799</id>
		<updated>2026-05-14T18:52:44Z</updated>
		<published>2026-04-22T23:50:10Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Python" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-179x100.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-300x168.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1536x862.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-500x280.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1024x574.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562.webp 1936w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Sparse-Deep-Learning" />In a previous post, we introduced the Universal Sparse Tensor (UST), enabling developers to decouple a tensor’s sparsity from its memory layout for greater...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/simplify-sparse-deep-learning-with-universal-sparse-tensor-in-nvmath-python/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-179x100.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-300x168.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1536x862.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-500x280.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1024x574.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562.webp 1936w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Sparse-Deep-Learning" />In a previous post, we introduced the Universal Sparse Tensor (UST), enabling developers to decouple a tensor’s sparsity from its memory layout for greater...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-179x100.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-300x168.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1536x862.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-500x280.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-1024x574.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Sparse-Deep-Learning-e1774302337562.webp 1936w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Sparse-Deep-Learning" /><p>In a previous post, we introduced the Universal Sparse Tensor (UST), enabling developers to decouple a tensor’s sparsity from its memory layout for greater flexibility and performance. We’re excited to announce the integration of the UST into nvmath-python v0.9.0 to accelerate sparse scientific and deep learning applications. This post provides a walkthrough of key UST features…</p>
<p><a href="https://developer.nvidia.com/blog/simplify-sparse-deep-learning-with-universal-sparse-tensor-in-nvmath-python/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/simplify-sparse-deep-learning-with-universal-sparse-tensor-in-nvmath-python/#comments" thr:count="3"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/simplify-sparse-deep-learning-with-universal-sparse-tensor-in-nvmath-python/feed/" thr:count="3"/>
		<thr:total>3</thr:total>
	</entry>
		<entry>
		<author>
			<name>Phoebe Lee</name>
					</author>
		<title type="html"><![CDATA[Scaling the AI-Ready Data Center with NVIDIA RTX PRO 4500 Blackwell Server Edition and NVIDIA vGPU 20]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/scaling-the-ai-ready-data-center-with-nvidia-rtx-pro-4500-blackwell-server-edition-and-nvidia-vgpu-20/" />
		<id>https://developer.nvidia.com/blog/?p=115767</id>
		<updated>2026-05-14T18:52:46Z</updated>
		<published>2026-04-22T20:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Multi-GPU" /><category scheme="https://developer.nvidia.com/blog" term="Multi-Instance GPU (MIG)" /><category scheme="https://developer.nvidia.com/blog" term="vGPU" /><category scheme="https://developer.nvidia.com/blog" term="Virtualization" /><category scheme="https://developer.nvidia.com/blog" term="VMware" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-blackwell-gpu" />AI integration is redefining mainstream enterprise applications, from productivity software like Microsoft Office to more complex design and engineering tools....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/scaling-the-ai-ready-data-center-with-nvidia-rtx-pro-4500-blackwell-server-edition-and-nvidia-vgpu-20/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-blackwell-gpu" />AI integration is redefining mainstream enterprise applications, from productivity software like Microsoft Office to more complex design and engineering tools....<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/nvidia-blackwell-gpu.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-blackwell-gpu" /><p>AI integration is redefining mainstream enterprise applications, from productivity software like Microsoft Office to more complex design and engineering tools. This shift requires the modern data center to move beyond single-purpose silos. For developers, gaining access to dedicated GPU compute can often be a bottleneck. Virtual machines (VMs) solve part of this challenge by providing secure…</p>
<p><a href="https://developer.nvidia.com/blog/scaling-the-ai-ready-data-center-with-nvidia-rtx-pro-4500-blackwell-server-edition-and-nvidia-vgpu-20/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/scaling-the-ai-ready-data-center-with-nvidia-rtx-pro-4500-blackwell-server-edition-and-nvidia-vgpu-20/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/scaling-the-ai-ready-data-center-with-nvidia-rtx-pro-4500-blackwell-server-edition-and-nvidia-vgpu-20/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Hao Wu</name>
					</author>
		<title type="html"><![CDATA[Advancing Emerging Optimizers for Accelerated LLM Training with NVIDIA Megatron]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/" />
		<id>https://developer.nvidia.com/blog/?p=115983</id>
		<updated>2026-05-14T18:52:47Z</updated>
		<published>2026-04-22T20:01:03Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB300" /><category scheme="https://developer.nvidia.com/blog" term="Megatron" /><category scheme="https://developer.nvidia.com/blog" term="NVIDIA Research" /><category scheme="https://developer.nvidia.com/blog" term="Training AI Models" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." />Higher-order optimization algorithms such as Shampoo have been effectively applied in neural network training for at least a decade. These methods have achieved...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." />Higher-order optimization algorithms such as Shampoo have been effectively applied in neural network training for at least a decade. These methods have achieved...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." /><p>Higher-order optimization algorithms such as Shampoo have been effectively applied in neural network training for at least a decade. These methods have achieved significant success more recently when applied to leading LLMs. In particular, Muon (MomentUm Orthogonalized by Newton-Schulz) was used to train some of today’s best open source models, including Kimi K2 and GLM-5.</p>
<p><a href="https://developer.nvidia.com/blog/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anshuman Bhat</name>
					</author>
		<title type="html"><![CDATA[Maximizing Memory Efficiency to Run Bigger Models on NVIDIA Jetson]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/maximizing-memory-efficiency-to-run-bigger-models-on-nvidia-jetson/" />
		<id>https://developer.nvidia.com/blog/?p=115920</id>
		<updated>2026-05-14T18:52:49Z</updated>
		<published>2026-04-20T23:01:04Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Robotics-Jetson-OSS" />The boom in open source generative AI models is pushing beyond data centers into machines operating in the physical world. Developers are eager to deploy these...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/maximizing-memory-efficiency-to-run-bigger-models-on-nvidia-jetson/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Robotics-Jetson-OSS" />The boom in open source generative AI models is pushing beyond data centers into machines operating in the physical world. Developers are eager to deploy these...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Robotics-Jetson-OSS.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Robotics-Jetson-OSS" /><p>The boom in open source generative AI models is pushing beyond data centers into machines operating in the physical world. Developers are eager to deploy these models at the edge, enabling physical AI agents and autonomous robots to automate heavy-duty tasks. A key challenge is efficiently running multi-billion-parameter models on edge devices with limited memory. With ongoing constraints on…</p>
<p><a href="https://developer.nvidia.com/blog/maximizing-memory-efficiency-to-run-bigger-models-on-nvidia-jetson/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/maximizing-memory-efficiency-to-run-bigger-models-on-nvidia-jetson/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/maximizing-memory-efficiency-to-run-bigger-models-on-nvidia-jetson/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Guyue Huang</name>
					</author>
		<title type="html"><![CDATA[Run High-Throughput Reinforcement Learning Training with End-to-End FP8 Precision]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precision/" />
		<id>https://developer.nvidia.com/blog/?p=115945</id>
		<updated>2026-05-14T18:52:51Z</updated>
		<published>2026-04-20T22:52:15Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Training AI Models" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922.webp 1173w" sizes="auto, (max-width: 768px) 100vw, 768px" title="RL-FP8" />As LLMs transition from simple text generation to complex reasoning, reinforcement learning (RL) plays a central role. Algorithms like Group Relative Policy...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precision/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922.webp 1173w" sizes="auto, (max-width: 768px) 100vw, 768px" title="RL-FP8" />As LLMs transition from simple text generation to complex reasoning, reinforcement learning (RL) plays a central role. Algorithms like Group Relative Policy...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/RL-FP8-e1776716106922.webp 1173w" sizes="auto, (max-width: 768px) 100vw, 768px" title="RL-FP8" /><p>As LLMs transition from simple text generation to complex reasoning, reinforcement learning (RL) plays a central role. Algorithms like Group Relative Policy Optimization (GRPO) power this transition, enabling reasoning-grade models to continuously improve through iterative feedback. Unlike standard supervised fine-tuning, RL training loops are bifurcated into two distinct, high-intensity phases: a…</p>
<p><a href="https://developer.nvidia.com/blog/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precision/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precision/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precision/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Daniel Teixeira</name>
					</author>
		<title type="html"><![CDATA[Mitigating Indirect AGENTS.md Injection Attacks in Agentic Environments]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/mitigating-indirect-agents-md-injection-attacks-in-agentic-environments/" />
		<id>https://developer.nvidia.com/blog/?p=115480</id>
		<updated>2026-05-14T18:52:53Z</updated>
		<published>2026-04-20T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI / Cybersecurity" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Red Team" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Security for AI" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" />AI tools are significantly accelerating software development and changing how developers work with code. These tools serve as real-time copilots, automating...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/mitigating-indirect-agents-md-injection-attacks-in-agentic-environments/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" />AI tools are significantly accelerating software development and changing how developers work with code. These tools serve as real-time copilots, automating...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/agentic-ai.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai" /><p>AI tools are significantly accelerating software development and changing how developers work with code. These tools serve as real-time copilots, automating repetitive tasks, executing tasks, writing documentation, and more. OpenAI Codex, for example, is a coding agent designed to assist developers through tasks like code generation, debugging, and automated pull request (PR) creation.</p>
<p><a href="https://developer.nvidia.com/blog/mitigating-indirect-agents-md-injection-attacks-in-agentic-environments/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/mitigating-indirect-agents-md-injection-attacks-in-agentic-environments/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/mitigating-indirect-agents-md-injection-attacks-in-agentic-environments/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ishan Dhanani</name>
					</author>
		<title type="html"><![CDATA[Full-Stack Optimizations for Agentic Inference with NVIDIA Dynamo]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/full-stack-optimizations-for-agentic-inference-with-nvidia-dynamo/" />
		<id>https://developer.nvidia.com/blog/?p=115673</id>
		<updated>2026-05-14T18:52:55Z</updated>
		<published>2026-04-17T22:52:47Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Build AI Agents" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" />Coding agents are starting to write production code at scale. Stripe’s agents generate 1,300+ PRs per week. Ramp attributes 30% of merged PRs to agents....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/full-stack-optimizations-for-agentic-inference-with-nvidia-dynamo/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" />Coding agents are starting to write production code at scale. Stripe’s agents generate 1,300+ PRs per week. Ramp attributes 30% of merged PRs to agents....<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" /><p>Coding agents are starting to write production code at scale. Stripe’s agents generate 1,300+ PRs per week. Ramp attributes 30% of merged PRs to agents. Spotify reports 650+ agent-generated PRs per month. Tools like Claude Code and Codex make hundreds of API calls per coding session, each carrying the full conversation history. Behind every one of these workflows is an inference stack under…</p>
<p><a href="https://developer.nvidia.com/blog/full-stack-optimizations-for-agentic-inference-with-nvidia-dynamo/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/full-stack-optimizations-for-agentic-inference-with-nvidia-dynamo/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/full-stack-optimizations-for-agentic-inference-with-nvidia-dynamo/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Patrick Moorhead</name>
					</author>
		<title type="html"><![CDATA[Build a More Secure, Always-On Local AI Agent with OpenClaw and NVIDIA NemoClaw]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-a-secure-always-on-local-ai-agent-with-nvidia-nemoclaw-and-openclaw/" />
		<id>https://developer.nvidia.com/blog/?p=115891</id>
		<updated>2026-04-30T17:41:58Z</updated>
		<published>2026-04-17T18:59:12Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI / Cybersecurity" /><category scheme="https://developer.nvidia.com/blog" term="Build AI Agents" /><category scheme="https://developer.nvidia.com/blog" term="claws" /><category scheme="https://developer.nvidia.com/blog" term="DGX Spark" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="NemoClaw" /><category scheme="https://developer.nvidia.com/blog" term="OpenShell" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Claw-DGX-Spark" />Agents are evolving from question-and-answer systems into long-running autonomous assistants that read files, call APIs, and drive multi-step workflows....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-a-secure-always-on-local-ai-agent-with-nvidia-nemoclaw-and-openclaw/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Claw-DGX-Spark" />Agents are evolving from question-and-answer systems into long-running autonomous assistants that read files, call APIs, and drive multi-step workflows....<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Claw-DGX-Spark.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Claw-DGX-Spark" /><p></p>
<p><a href="https://developer.nvidia.com/blog/build-a-secure-always-on-local-ai-agent-with-nvidia-nemoclaw-and-openclaw/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-a-secure-always-on-local-ai-agent-with-nvidia-nemoclaw-and-openclaw/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-a-secure-always-on-local-ai-agent-with-nvidia-nemoclaw-and-openclaw/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Mark Hobbs</name>
					</author>
		<title type="html"><![CDATA[Accelerate Clean, Modular, Nuclear Reactor Design with AI Physics]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/accelerate-clean-modular-nuclear-reactor-design-with-ai-physics/" />
		<id>https://developer.nvidia.com/blog/?p=115829</id>
		<updated>2026-04-30T17:41:59Z</updated>
		<published>2026-04-17T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Energy" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="PhysicsNeMo" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="image8" />The development of socially acceptable nuclear reactors requires that they are safe, clean, efficient, economical, and sustainable. Meeting these requirements...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/accelerate-clean-modular-nuclear-reactor-design-with-ai-physics/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="image8" />The development of socially acceptable nuclear reactors requires that they are safe, clean, efficient, economical, and sustainable. Meeting these requirements...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image8-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="image8" /><p>The development of socially acceptable nuclear reactors requires that they are safe, clean, efficient, economical, and sustainable. Meeting these requirements calls for new approaches, driving growing interest in Small Modular Reactors (SMRs) and in Generation IV designs. SMRs aim to improve project economics by standardising designs and shifting construction to controlled manufacturing…</p>
<p><a href="https://developer.nvidia.com/blog/accelerate-clean-modular-nuclear-reactor-design-with-ai-physics/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/accelerate-clean-modular-nuclear-reactor-design-with-ai-physics/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/accelerate-clean-modular-nuclear-reactor-design-with-ai-physics/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Debraj Sinha</name>
					</author>
		<title type="html"><![CDATA[How to Build Vision AI Pipelines Using NVIDIA DeepStream Coding Agents ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-vision-ai-pipelines-using-deepstream-coding-agents/" />
		<id>https://developer.nvidia.com/blog/?p=115804</id>
		<updated>2026-04-30T17:42:00Z</updated>
		<published>2026-04-16T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Computer Vision / Video Analytics" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="DeepStream" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Metropolis" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080" />Developing real-time vision AI applications presents a significant challenge for developers, often demanding intricate data pipelines, countless lines of code,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-build-vision-ai-pipelines-using-deepstream-coding-agents/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080" />Developing real-time vision AI applications presents a significant challenge for developers, often demanding intricate data pipelines, countless lines of code,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotics-key-visual-metropolis-deepstream-gtc26-abstract-code-r2_1920x1080" /><p>Developing real-time vision AI applications presents a significant challenge for developers, often demanding intricate data pipelines, countless lines of code, and lengthy development cycles. NVIDIA DeepStream 9 removes these development barriers using coding agents, such as Claude Code or Cursor, to help you easily create deployable, optimized code that brings your vision AI applications to…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-build-vision-ai-pipelines-using-deepstream-coding-agents/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-vision-ai-pipelines-using-deepstream-coding-agents/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-build-vision-ai-pipelines-using-deepstream-coding-agents/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Erica Tsai</name>
					</author>
		<title type="html"><![CDATA[Building Custom Atomistic Simulation Workflows for Chemistry and Materials Science with NVIDIA ALCHEMI Toolkit]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-custom-atomistic-simulation-workflows-for-chemistry-and-materials-science-with-nvidia-alchemi-toolkit/" />
		<id>https://developer.nvidia.com/blog/?p=115414</id>
		<updated>2026-04-30T17:42:02Z</updated>
		<published>2026-04-14T16:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="ALCHEMI" /><category scheme="https://developer.nvidia.com/blog" term="Computational Chemistry / Materials Science" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="PyTorch" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="materials-science-chemistry" />For decades, computational chemistry has faced a tug-of-war between accuracy and speed. Ab initio methods like density functional theory (DFT) provide high...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-custom-atomistic-simulation-workflows-for-chemistry-and-materials-science-with-nvidia-alchemi-toolkit/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="materials-science-chemistry" />For decades, computational chemistry has faced a tug-of-war between accuracy and speed. Ab initio methods like density functional theory (DFT) provide high...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/materials-science-chemistry-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="materials-science-chemistry" /><p>For decades, computational chemistry has faced a tug-of-war between accuracy and speed. Ab initio methods like density functional theory (DFT) provide high fidelity but are computationally expensive, limiting researchers to systems of a few hundred atoms. Conversely, classical force fields are fast but often lack the chemical accuracy required for complex bond-breaking or transition-state analysis.</p>
<p><a href="https://developer.nvidia.com/blog/building-custom-atomistic-simulation-workflows-for-chemistry-and-materials-science-with-nvidia-alchemi-toolkit/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-custom-atomistic-simulation-workflows-for-chemistry-and-materials-science-with-nvidia-alchemi-toolkit/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-custom-atomistic-simulation-workflows-for-chemistry-and-materials-science-with-nvidia-alchemi-toolkit/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Eva Sitaridi</name>
					</author>
		<title type="html"><![CDATA[NVIDIA NVbandwidth: Your Essential Tool for Measuring GPU Interconnect and Memory Performance]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-nvbandwidth-your-essential-tool-for-measuring-gpu-interconnect-and-memory-performance/" />
		<id>https://developer.nvidia.com/blog/?p=115566</id>
		<updated>2026-04-30T17:42:03Z</updated>
		<published>2026-04-14T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="CUDA" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="NCCL" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-625x351.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-645x362.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-362x203.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-1024x575.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-jpg.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" />When you’re writing CUDA applications, one of the most important things you need to focus on to write great code is data transfer performance. This applies to...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-nvbandwidth-your-essential-tool-for-measuring-gpu-interconnect-and-memory-performance/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-625x351.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-645x362.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-362x203.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-1024x575.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-jpg.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" />When you’re writing CUDA applications, one of the most important things you need to focus on to write great code is data transfer performance. This applies to...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-625x351.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-645x362.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-362x203.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-1024x575.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/image1-1-jpg.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" /><p>When you’re writing CUDA applications, one of the most important things you need to focus on to write great code is data transfer performance. This applies to both single-GPU and multi-GPU systems alike. One of the tools you can use to understand the memory characteristics of your GPU system is NVIDIA NVbandwidth. In this blog post, we’ll explore what NVbandwidth is, how it works…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-nvbandwidth-your-essential-tool-for-measuring-gpu-interconnect-and-memory-performance/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-nvbandwidth-your-essential-tool-for-measuring-gpu-interconnect-and-memory-performance/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-nvbandwidth-your-essential-tool-for-measuring-gpu-interconnect-and-memory-performance/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Tom Lubowe</name>
					</author>
		<title type="html"><![CDATA[NVIDIA Ising Introduces AI-Powered Workflows to Build Fault-Tolerant Quantum Systems]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-ising-introduces-ai-powered-workflows-to-build-fault-tolerant-quantum-systems/" />
		<id>https://developer.nvidia.com/blog/?p=115554</id>
		<updated>2026-04-30T17:42:04Z</updated>
		<published>2026-04-14T14:15:56Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Ising" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Ising-Quantum" />NVIDIA Ising is the world's first family of open AI models for building quantum processors, launching with two model domains: Ising Calibration and Ising...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-ising-introduces-ai-powered-workflows-to-build-fault-tolerant-quantum-systems/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Ising-Quantum" />NVIDIA Ising is the world's first family of open AI models for building quantum processors, launching with two model domains: Ising Calibration and Ising...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Ising-Quantum.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Ising-Quantum" /><p>NVIDIA Ising is the world’s first family of open AI models for building quantum processors, launching with two model domains: Ising Calibration and Ising Decoding. Both target the fundamental challenge in quantum computing—qubits are inherently noisy. The best quantum processors make an error roughly once in every thousand operations. To become useful accelerators for scientific and…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-ising-introduces-ai-powered-workflows-to-build-fault-tolerant-quantum-systems/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-ising-introduces-ai-powered-workflows-to-build-fault-tolerant-quantum-systems/#comments" thr:count="3"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-ising-introduces-ai-powered-workflows-to-build-fault-tolerant-quantum-systems/feed/" thr:count="3"/>
		<thr:total>3</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anu Srivastava</name>
					</author>
		<title type="html"><![CDATA[MiniMax M2.7 Advances Scalable Agentic Workflows on NVIDIA Platforms for Complex AI Applications ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/minimax-m2-7-advances-scalable-agentic-workflows-on-nvidia-platforms-for-complex-ai-applications/" />
		<id>https://developer.nvidia.com/blog/?p=115559</id>
		<updated>2026-04-30T17:42:06Z</updated>
		<published>2026-04-12T01:02:44Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Mixture of Experts (MoE)" /><category scheme="https://developer.nvidia.com/blog" term="NemoClaw" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative object." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MM-Release" />The release of&nbsp;MiniMax M2.7&nbsp;adds enhancements to the popular&nbsp;MiniMax&nbsp;M2.5 model,&nbsp;built for agentic harnesses,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/minimax-m2-7-advances-scalable-agentic-workflows-on-nvidia-platforms-for-complex-ai-applications/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative object." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MM-Release" />The release of&nbsp;MiniMax M2.7&nbsp;adds enhancements to the popular&nbsp;MiniMax&nbsp;M2.5 model,&nbsp;built for agentic harnesses,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative object." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/MM-Release.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="MM-Release" /><p>The release of MiniMax M2.7 adds enhancements to the popular MiniMax M2.5 model, built for agentic harnesses, and other complex use cases in fields such as reasoning, ML research workflows, software, engineering, and office work. The open weights release of MiniMax M2.7 is now available through NVIDIA and across the open source inference ecosystem. The MiniMax M2 series is a sparse mixture-of…</p>
<p><a href="https://developer.nvidia.com/blog/minimax-m2-7-advances-scalable-agentic-workflows-on-nvidia-platforms-for-complex-ai-applications/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/minimax-m2-7-advances-scalable-agentic-workflows-on-nvidia-platforms-for-complex-ai-applications/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/minimax-m2-7-advances-scalable-agentic-workflows-on-nvidia-platforms-for-complex-ai-applications/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anton Polyakov</name>
					</author>
		<title type="html"><![CDATA[Running Large-Scale GPU Workloads on Kubernetes with Slurm]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/running-large-scale-gpu-workloads-on-kubernetes-with-slurm/" />
		<id>https://developer.nvidia.com/blog/?p=115345</id>
		<updated>2026-05-07T18:09:16Z</updated>
		<published>2026-04-09T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Cloud Services" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Kubernetes" /><category scheme="https://developer.nvidia.com/blog" term="Slurm" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack.webp 1921w" sizes="auto, (max-width: 768px) 100vw, 768px" title="compute-stack" />Slurm is an open source cluster management and job scheduling system for Linux. It manages job scheduling for over 65% of TOP500 systems. Most organizations...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/running-large-scale-gpu-workloads-on-kubernetes-with-slurm/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack.webp 1921w" sizes="auto, (max-width: 768px) 100vw, 768px" title="compute-stack" />Slurm is an open source cluster management and job scheduling system for Linux. It manages job scheduling for over 65% of TOP500 systems. Most organizations...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/compute-stack.webp 1921w" sizes="auto, (max-width: 768px) 100vw, 768px" title="compute-stack" /><p>Slurm is an open source cluster management and job scheduling system for Linux. It manages job scheduling for over 65% of TOP500 systems. Most organizations running large-scale AI training have years of investment in Slurm job scripts, fair-share policies, and accounting workflows. The challenge is getting Slurm scheduling capabilities onto Kubernetes—the standard platform for managing GPU…</p>
<p><a href="https://developer.nvidia.com/blog/running-large-scale-gpu-workloads-on-kubernetes-with-slurm/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/running-large-scale-gpu-workloads-on-kubernetes-with-slurm/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/running-large-scale-gpu-workloads-on-kubernetes-with-slurm/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Wenqi Glantz</name>
					</author>
		<title type="html"><![CDATA[Cut Checkpoint Costs with About 30 Lines of Python and NVIDIA nvCOMP]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/cut-checkpoint-costs-with-about-30-lines-of-python-and-nvidia-nvcomp/" />
		<id>https://developer.nvidia.com/blog/?p=115453</id>
		<updated>2026-04-30T17:42:08Z</updated>
		<published>2026-04-09T16:48:38Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565.webp 1843w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Checkpoint-Costs" />Training LLMs requires periodic checkpoints. These full snapshots of model weights, optimizer states, and gradients are saved to storage so training can resume...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/cut-checkpoint-costs-with-about-30-lines-of-python-and-nvidia-nvcomp/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565.webp 1843w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Checkpoint-Costs" />Training LLMs requires periodic checkpoints. These full snapshots of model weights, optimizer states, and gradients are saved to storage so training can resume...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Checkpoint-Costs-e1775685819565.webp 1843w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Checkpoint-Costs" /><p>Training LLMs requires periodic checkpoints. These full snapshots of model weights, optimizer states, and gradients are saved to storage so training can resume after interruptions. At scale, these checkpoints become massive (782 GB for a 70B model) and frequent (every 15-30 minutes), generating one of the largest line items in a training budget. Most AI teams chase GPU utilization…</p>
<p><a href="https://developer.nvidia.com/blog/cut-checkpoint-costs-with-about-30-lines-of-python-and-nvidia-nvcomp/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/cut-checkpoint-costs-with-about-30-lines-of-python-and-nvidia-nvcomp/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/cut-checkpoint-costs-with-about-30-lines-of-python-and-nvidia-nvcomp/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Christian Dallago</name>
					</author>
		<title type="html"><![CDATA[How to Accelerate Protein Structure Prediction at Proteome-Scale]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-accelerate-protein-structure-prediction-at-proteome-scale/" />
		<id>https://developer.nvidia.com/blog/?p=115196</id>
		<updated>2026-04-30T17:42:09Z</updated>
		<published>2026-04-09T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="BioNeMo" /><category scheme="https://developer.nvidia.com/blog" term="CUDA" /><category scheme="https://developer.nvidia.com/blog" term="Drug Discovery" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Healthcare &amp; Life Sciences" /><category scheme="https://developer.nvidia.com/blog" term="HPC / Scientific Computing" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5.webp 1930w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" />Proteins rarely function in isolation as individual monomers. Most biological processes are governed by proteins interacting with other proteins, forming...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-accelerate-protein-structure-prediction-at-proteome-scale/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5.webp 1930w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" />Proteins rarely function in isolation as individual monomers. Most biological processes are governed by proteins interacting with other proteins, forming...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/image1-5.webp 1930w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image1" /><p>Proteins rarely function in isolation as individual monomers. Most biological processes are governed by proteins interacting with other proteins, forming protein complexes whose structures are described in the hierarchy of protein structure as the quaternary representation. This represents one level of complexity up from tertiary representations, the 3D structure of monomers…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-accelerate-protein-structure-prediction-at-proteome-scale/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-accelerate-protein-structure-prediction-at-proteome-scale/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-accelerate-protein-structure-prediction-at-proteome-scale/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ashley Goldstein</name>
					</author>
		<title type="html"><![CDATA[Integrate Physical AI Capabilities into Existing Apps with NVIDIA Omniverse Libraries]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/integrate-physical-ai-capabilities-into-existing-apps-with-nvidia-omniverse-libraries/" />
		<id>https://developer.nvidia.com/blog/?p=115313</id>
		<updated>2026-04-30T17:42:10Z</updated>
		<published>2026-04-08T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Graphic Design" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Industrial Digitalization / Digital Twin" /><category scheme="https://developer.nvidia.com/blog" term="Isaac Lab" /><category scheme="https://developer.nvidia.com/blog" term="Isaac Sim" /><category scheme="https://developer.nvidia.com/blog" term="Omniverse" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-libraries-tech-blog-1920x1080" />Physical AI—AI systems that perceive, reason, and act in physically grounded simulated environments—is changing how teams design and validate robots and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/integrate-physical-ai-capabilities-into-existing-apps-with-nvidia-omniverse-libraries/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-libraries-tech-blog-1920x1080" />Physical AI—AI systems that perceive, reason, and act in physically grounded simulated environments—is changing how teams design and validate robots and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/ov-libraries-tech-blog-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-libraries-tech-blog-1920x1080" /><p>Physical AI—AI systems that perceive, reason, and act in physically grounded simulated environments—is changing how teams design and validate robots and industrial systems, long before anything ships to the factory floor. At GTC 2026, NVIDIA highlighted physical AI as a key direction for robotics and digital twins, where policies are trained and validated against physically grounded environments.</p>
<p><a href="https://developer.nvidia.com/blog/integrate-physical-ai-capabilities-into-existing-apps-with-nvidia-omniverse-libraries/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/integrate-physical-ai-capabilities-into-existing-apps-with-nvidia-omniverse-libraries/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/integrate-physical-ai-capabilities-into-existing-apps-with-nvidia-omniverse-libraries/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ryan Prout</name>
					</author>
		<title type="html"><![CDATA[Running AI Workloads on Rack-Scale Supercomputers: From Hardware to Topology-Aware Scheduling]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/running-ai-workloads-on-rack-scale-supercomputers-from-hardware-to-topology-aware-scheduling/" />
		<id>https://developer.nvidia.com/blog/?p=114998</id>
		<updated>2026-04-30T17:42:11Z</updated>
		<published>2026-04-07T18:51:01Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-gb300" />The NVIDIA GB200 NVL72 and NVIDIA GB300 NVL72 systems, featuring NVIDIA Blackwell architecture, are rack-scale supercomputers. They’re designed with 18...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/running-ai-workloads-on-rack-scale-supercomputers-from-hardware-to-topology-aware-scheduling/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-gb300" />The NVIDIA GB200 NVL72 and NVIDIA GB300 NVL72 systems, featuring NVIDIA Blackwell architecture, are rack-scale supercomputers. They’re designed with 18...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/gtc25-tech-blog-dgx-gb300-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-gb300" /><p>The NVIDIA GB200 NVL72 and NVIDIA GB300 NVL72 systems, featuring NVIDIA Blackwell architecture, are rack-scale supercomputers. They’re designed with 18 tightly coupled compute trays, massive GPU fabrics, and high-bandwidth networking packaged as a unit. For AI architects and HPC platform operators, the challenge isn’t just racking and stacking hardware—it’s turning infrastructure into safe…</p>
<p><a href="https://developer.nvidia.com/blog/running-ai-workloads-on-rack-scale-supercomputers-from-hardware-to-topology-aware-scheduling/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/running-ai-workloads-on-rack-scale-supercomputers-from-hardware-to-topology-aware-scheduling/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/running-ai-workloads-on-rack-scale-supercomputers-from-hardware-to-topology-aware-scheduling/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Andreas Kieslinger</name>
					</author>
		<title type="html"><![CDATA[Accelerating Vision AI Pipelines with Batch Mode VC-6 and NVIDIA Nsight]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/accelerating-vision-ai-pipelines-with-batch-mode-vc-6-and-nvidia-nsight/" />
		<id>https://developer.nvidia.com/blog/?p=115141</id>
		<updated>2026-04-16T17:15:09Z</updated>
		<published>2026-04-02T20:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI workflows" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Video Decode / Encode" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-2048x1152-jpg.webp 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-960x540-jpg.webp 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="blue-square-field" />In vision AI systems, model throughput continues to improve. The surrounding pipeline stages must keep pace, including decode, preprocessing, and GPU...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/accelerating-vision-ai-pipelines-with-batch-mode-vc-6-and-nvidia-nsight/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-2048x1152-jpg.webp 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-960x540-jpg.webp 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="blue-square-field" />In vision AI systems, model throughput continues to improve. The surrounding pipeline stages must keep pace, including decode, preprocessing, and GPU...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-768x432-jpg.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-300x169-jpg.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-625x352-jpg.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-179x101-jpg.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1536x864-jpg.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-2048x1152-jpg.webp 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-645x363-jpg.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-660x370-jpg.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-500x281-jpg.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-160x90-jpg.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-362x204-jpg.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-195x110-jpg.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-1024x576-jpg.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/08/blue-square-field-960x540-jpg.webp 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="blue-square-field" /><p>In vision AI systems, model throughput continues to improve. The surrounding pipeline stages must keep pace, including decode, preprocessing, and GPU scheduling. In the previous post, Build High-Performance Vision AI Pipelines with NVIDIA CUDA-Accelerated VC-6, this was described as the data-to-tensor gap—a performance mismatch between AI pipeline stages. The SMPTE VC-6 (ST 2117-1) codec…</p>
<p><a href="https://developer.nvidia.com/blog/accelerating-vision-ai-pipelines-with-batch-mode-vc-6-and-nvidia-nsight/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/accelerating-vision-ai-pipelines-with-batch-mode-vc-6-and-nvidia-nsight/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/accelerating-vision-ai-pipelines-with-batch-mode-vc-6-and-nvidia-nsight/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anu Srivastava</name>
					</author>
		<title type="html"><![CDATA[Bringing AI Closer to the Edge and On-Device with Gemma 4 ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/bringing-ai-closer-to-the-edge-and-on-device-with-gemma-4/" />
		<id>https://developer.nvidia.com/blog/?p=115165</id>
		<updated>2026-04-16T17:15:11Z</updated>
		<published>2026-04-02T16:27:46Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured.webp 2003w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Gemma 4 TechBlog featured" />The Gemmaverse expands with the launch of the latest Gemma 4 multimodal and multilingual models, designed to scale across the full spectrum of deployments, from...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/bringing-ai-closer-to-the-edge-and-on-device-with-gemma-4/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured.webp 2003w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Gemma 4 TechBlog featured" />The Gemmaverse expands with the launch of the latest Gemma 4 multimodal and multilingual models, designed to scale across the full spectrum of deployments, from...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/Gemma-4-TechBlog-featured.webp 2003w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Gemma 4 TechBlog featured" /><p>The Gemmaverse expands with the launch of the latest Gemma 4 multimodal and multilingual models, designed to scale across the full spectrum of deployments, from NVIDIA Blackwell in the data center to Jetson at the edge. These models are suited to meet the growing demand for local deployment for AI development and prototyping, secure on-prem requirements, cost efficiency, and latency-sensitive use…</p>
<p><a href="https://developer.nvidia.com/blog/bringing-ai-closer-to-the-edge-and-on-device-with-gemma-4/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
<link href="https://developer-blogs.nvidia.com/wp-content/uploads/2026/04/EeveeDemo.mov" rel="enclosure" length="7959691" type="video/quicktime" />
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/bringing-ai-closer-to-the-edge-and-on-device-with-gemma-4/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/bringing-ai-closer-to-the-edge-and-on-device-with-gemma-4/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Nikolay Markovskiy</name>
					</author>
		<title type="html"><![CDATA[Achieving Single-Digit Microsecond Latency Inference for Capital Markets]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/achieving-single-digit-microsecond-latency-inference-for-capital-markets/" />
		<id>https://developer.nvidia.com/blog/?p=115102</id>
		<updated>2026-04-16T17:15:13Z</updated>
		<published>2026-04-02T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Low-Latency Inference" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="finance-trading" />In algorithmic trading, reducing response times to market events is crucial. To keep pace with high-speed electronic markets, latency-sensitive firms often use...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/achieving-single-digit-microsecond-latency-inference-for-capital-markets/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="finance-trading" />In algorithmic trading, reducing response times to market events is crucial. To keep pace with high-speed electronic markets, latency-sensitive firms often use...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/finance-trading.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="finance-trading" /><p>In algorithmic trading, reducing response times to market events is crucial. To keep pace with high-speed electronic markets, latency-sensitive firms often use specialized hardware like FPGAs and ASICs. Yet, as markets grow more efficient, traders increasingly depend on advanced models such as deep neural networks to enhance profitability. Because implementing these complex models on low-level…</p>
<p><a href="https://developer.nvidia.com/blog/achieving-single-digit-microsecond-latency-inference-for-capital-markets/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/achieving-single-digit-microsecond-latency-inference-for-capital-markets/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/achieving-single-digit-microsecond-latency-inference-for-capital-markets/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Rob Armstrong</name>
					</author>
		<title type="html"><![CDATA[CUDA Tile Programming Now Available for BASIC!]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/cuda-tile-programming-now-available-for-basic/" />
		<id>https://developer.nvidia.com/blog/?p=115121</id>
		<updated>2026-04-16T17:15:14Z</updated>
		<published>2026-04-01T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="April Fools" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="News" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-625x351.gif 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-645x363.gif 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-660x370.gif 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-500x281.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-196x110.gif 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-1024x576.gif 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-960x540.gif 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1.gif 1138w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA Tile BASIC" />Note: CUDA Tile Programming in BASIC is an April Fools’ joke, but it's also real and actually works,  demonstrating the flexibility of CUDA. CUDA 13.1...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/cuda-tile-programming-now-available-for-basic/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-625x351.gif 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-645x363.gif 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-660x370.gif 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-500x281.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-196x110.gif 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-1024x576.gif 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-960x540.gif 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1.gif 1138w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA Tile BASIC" />Note: CUDA Tile Programming in BASIC is an April Fools’ joke, but it's also real and actually works,  demonstrating the flexibility of CUDA. CUDA 13.1...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-768x432.gif 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-625x351.gif 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-645x363.gif 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-660x370.gif 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-500x281.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-196x110.gif 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-1024x576.gif 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1-960x540.gif 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/0331-1.gif 1138w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA Tile BASIC" /><p>Note: CUDA Tile Programming in BASIC is an April Fools’ joke, but it’s also real and actually works, demonstrating the flexibility of CUDA. CUDA 13.1 introduced CUDA Tile, a next generation tile-based GPU programming paradigm designed to make fine-grained parallelism more accessible and flexible. One of its key strengths is language openness: any programming language can target CUDA Tile…</p>
<p><a href="https://developer.nvidia.com/blog/cuda-tile-programming-now-available-for-basic/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/cuda-tile-programming-now-available-for-basic/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/cuda-tile-programming-now-available-for-basic/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ashraf Eassa</name>
					</author>
		<title type="html"><![CDATA[NVIDIA Platform Delivers Lowest Token Cost Enabled by Extreme Co-Design]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-platform-delivers-lowest-token-cost-enabled-by-extreme-co-design/" />
		<id>https://developer.nvidia.com/blog/?p=115040</id>
		<updated>2026-04-16T17:15:16Z</updated>
		<published>2026-04-01T15:00:48Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell Ultra" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB300" /><category scheme="https://developer.nvidia.com/blog" term="Groq 3 LPX" /><category scheme="https://developer.nvidia.com/blog" term="Rubin GPU" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVL72" />Co-designed hardware, software, and models are key to delivering the highest AI factory throughput and lowest token cost. Measuring this goes far beyond peak...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-platform-delivers-lowest-token-cost-enabled-by-extreme-co-design/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVL72" />Co-designed hardware, software, and models are key to delivering the highest AI factory throughput and lowest token cost. Measuring this goes far beyond peak...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/NVL72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="NVL72" /><p>Co-designed hardware, software, and models are key to delivering the highest AI factory throughput and lowest token cost. Measuring this goes far beyond peak chip specifications. Rigorous AI inference performance benchmarks are critical to understanding real-world token output, which drives AI factory revenue. MLPerf Inference v6.0 is the latest in a series of industry benchmarks that measure…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-platform-delivers-lowest-token-cost-enabled-by-extreme-co-design/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-platform-delivers-lowest-token-cost-enabled-by-extreme-co-design/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-platform-delivers-lowest-token-cost-enabled-by-extreme-co-design/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Pradyumna Desale</name>
					</author>
		<title type="html"><![CDATA[Accelerate Token Production in AI Factories Using Unified Services and Real-Time AI]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/accelerate-token-production-in-ai-factories-using-unified-services-and-real-time-ai/" />
		<id>https://developer.nvidia.com/blog/?p=114947</id>
		<updated>2026-04-16T17:15:17Z</updated>
		<published>2026-04-01T15:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Data Platform" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="AI Platform" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell" /><category scheme="https://developer.nvidia.com/blog" term="Cloud Services" /><category scheme="https://developer.nvidia.com/blog" term="DGX" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />In today’s AI factory environment, performance is not theoretical. It is economic, competitive, and existential. A 1% drop in usable GPU time can mean...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/accelerate-token-production-in-ai-factories-using-unified-services-and-real-time-ai/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />In today’s AI factory environment, performance is not theoretical. It is economic, competitive, and existential. A 1% drop in usable GPU time can mean...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-9.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" /><p>In today’s AI factory environment, performance is not theoretical. It is economic, competitive, and existential. A 1% drop in usable GPU time can mean millions of tokens lost per hour. Minutes of congestion can cascade into hours of recovery. A rack-level power oversubscription can lead to stranded power and reduced tokens per watt, silently eroding factory output at scale. As AI factories scale…</p>
<p><a href="https://developer.nvidia.com/blog/accelerate-token-production-in-ai-factories-using-unified-services-and-real-time-ai/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/accelerate-token-production-in-ai-factories-using-unified-services-and-real-time-ai/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/accelerate-token-production-in-ai-factories-using-unified-services-and-real-time-ai/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Max Bickley</name>
					</author>
		<title type="html"><![CDATA[Stream High-Fidelity Spatial Computing Content to Any Device with NVIDIA CloudXR 6.0]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/stream-high-fidelity-spatial-computing-content-to-any-device-with-nvidia-cloudxr-6-0/" />
		<id>https://developer.nvidia.com/blog/?p=114702</id>
		<updated>2026-04-16T17:15:19Z</updated>
		<published>2026-03-31T18:14:50Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="AR / VR" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Visualization" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Workstation-CloudXr" />Spatial computing is moving from visualization to active collaboration, adding increasingly more GPU demands on XR hardware to render photorealistic,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/stream-high-fidelity-spatial-computing-content-to-any-device-with-nvidia-cloudxr-6-0/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Workstation-CloudXr" />Spatial computing is moving from visualization to active collaboration, adding increasingly more GPU demands on XR hardware to render photorealistic,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Workstation-CloudXr.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Workstation-CloudXr" /><p>Spatial computing is moving from visualization to active collaboration, adding increasingly more GPU demands on XR hardware to render photorealistic, physics-accurate, high-fidelity spatial content in real time. Meanwhile, developers have had to maintain separate codebases for every platform, each with different toolchains, SDKs, and streaming protocols. At NVIDIA GTC 2026, NVIDIA CloudXR 6.0…</p>
<p><a href="https://developer.nvidia.com/blog/stream-high-fidelity-spatial-computing-content-to-any-device-with-nvidia-cloudxr-6-0/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/stream-high-fidelity-spatial-computing-content-to-any-device-with-nvidia-cloudxr-6-0/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/stream-high-fidelity-spatial-computing-content-to-any-device-with-nvidia-cloudxr-6-0/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Yanzi Zhu</name>
					</author>
		<title type="html"><![CDATA[Build and Stream Browser-Based XR Experiences with NVIDIA CloudXR.js]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-and-stream-browser-based-xr-experiences-with-nvidia-cloudxr-js/" />
		<id>https://developer.nvidia.com/blog/?p=114958</id>
		<updated>2026-04-16T17:15:21Z</updated>
		<published>2026-03-31T17:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="AR / VR" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="CloudXR" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Industrial Digitalization / Digital Twin" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1.webp 1862w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotic-assembly-line" />Delivering high-fidelity VR and AR experiences to enterprise users has typically required native application development, custom device management, and complex...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-and-stream-browser-based-xr-experiences-with-nvidia-cloudxr-js/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1.webp 1862w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotic-assembly-line" />Delivering high-fidelity VR and AR experiences to enterprise users has typically required native application development, custom device management, and complex...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/robotic-assembly-line-1.webp 1862w" sizes="auto, (max-width: 768px) 100vw, 768px" title="robotic-assembly-line" /><p>Delivering high-fidelity VR and AR experiences to enterprise users has typically required native application development, custom device management, and complex deployment pipelines. Now, with the new JavaScript SDK NVIDIA CloudXR.js, developers can stream GPU-rendered immersive content directly to a standard web browser—no app store, no installs, no device-specific builds. NVIDIA CloudXR.</p>
<p><a href="https://developer.nvidia.com/blog/build-and-stream-browser-based-xr-experiences-with-nvidia-cloudxr-js/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-and-stream-browser-based-xr-experiences-with-nvidia-cloudxr-js/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-and-stream-browser-based-xr-experiences-with-nvidia-cloudxr-js/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Sagar Desai</name>
					</author>
		<title type="html"><![CDATA[Maximize AI Infrastructure Throughput by Consolidating Underutilized GPU Workloads]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/maximize-ai-infrastructure-throughput-by-consolidating-underutilized-gpu-workloads/" />
		<id>https://developer.nvidia.com/blog/?p=114752</id>
		<updated>2026-04-16T17:15:22Z</updated>
		<published>2026-03-25T16:35:43Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-657x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242.webp 1800w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Close-up of NVIDIA processors on a server motherboard." />In production Kubernetes environments, the difference between model requirements and GPU size creates inefficiencies. Lightweight automatic speech recognition...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/maximize-ai-infrastructure-throughput-by-consolidating-underutilized-gpu-workloads/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-657x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242.webp 1800w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Close-up of NVIDIA processors on a server motherboard." />In production Kubernetes environments, the difference between model requirements and GPU size creates inefficiencies. Lightweight automatic speech recognition...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-657x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/CUDA-MPS-e1765825617242.webp 1800w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Close-up of NVIDIA processors on a server motherboard." /><p>In production Kubernetes environments, the difference between model requirements and GPU size creates inefficiencies. Lightweight automatic speech recognition (ASR) or text-to-speech (TTS) models may require only 10 GB of VRAM, yet occupy an entire GPU in standard Kubernetes deployments. Because the scheduler maps a model to one or more GPUs and can’t easily share across GPUs across models…</p>
<p><a href="https://developer.nvidia.com/blog/maximize-ai-infrastructure-throughput-by-consolidating-underutilized-gpu-workloads/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/maximize-ai-infrastructure-throughput-by-consolidating-underutilized-gpu-workloads/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/maximize-ai-infrastructure-throughput-by-consolidating-underutilized-gpu-workloads/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Lachlan Dowling</name>
					</author>
		<title type="html"><![CDATA[How Centralized Radar Processing on NVIDIA DRIVE Enables Safer, Smarter Level 4 Autonomy]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-centralized-radar-processing-on-nvidia-drive-enables-safer-smarter-level-4-autonomy/" />
		<id>https://developer.nvidia.com/blog/?p=114855</id>
		<updated>2026-04-16T17:15:23Z</updated>
		<published>2026-03-25T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Computer Graphics &amp; Visualization" /><category scheme="https://developer.nvidia.com/blog" term="DRIVE" /><category scheme="https://developer.nvidia.com/blog" term="DRIVE AGX" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Lidar" /><category scheme="https://developer.nvidia.com/blog" term="Radar" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Feature1_new" />In the current state of automotive radar, machine learning engineers can't work with camera-equivalent raw RGB images. Instead, they work with the output of...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-centralized-radar-processing-on-nvidia-drive-enables-safer-smarter-level-4-autonomy/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Feature1_new" />In the current state of automotive radar, machine learning engineers can't work with camera-equivalent raw RGB images. Instead, they work with the output of...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Feature1_new-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Feature1_new" /><p>In the current state of automotive radar, machine learning engineers can’t work with camera-equivalent raw RGB images. Instead, they work with the output of radar constant false alarm rate (CFAR), which is similar to computer vision (CV) edge detections. The communications and compute architectures haven’t kept pace with trends in AI and the needs of Level 4 autonomy, despite radar being a staple…</p>
<p><a href="https://developer.nvidia.com/blog/how-centralized-radar-processing-on-nvidia-drive-enables-safer-smarter-level-4-autonomy/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
<link href="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Video1-1.mp4" rel="enclosure" length="6205963" type="video/mp4" />
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-centralized-radar-processing-on-nvidia-drive-enables-safer-smarter-level-4-autonomy/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-centralized-radar-processing-on-nvidia-drive-enables-safer-smarter-level-4-autonomy/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Kyle Gion</name>
					</author>
		<title type="html"><![CDATA[Designing Protein Binders Using the Generative Model Proteina-Complexa]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/designing-protein-binders-using-the-generative-model-proteina-complexa/" />
		<id>https://developer.nvidia.com/blog/?p=114669</id>
		<updated>2026-04-16T17:15:25Z</updated>
		<published>2026-03-25T13:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1536x865.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-500x282.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1024x577.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Proteina-Complexa-Binder-1920x1080" />Developing new protein-based therapies and catalysts involves the challenging task of designing protein binders, or proteins that bind to a target protein or...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/designing-protein-binders-using-the-generative-model-proteina-complexa/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1536x865.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-500x282.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1024x577.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Proteina-Complexa-Binder-1920x1080" />Developing new protein-based therapies and catalysts involves the challenging task of designing protein binders, or proteins that bind to a target protein or...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1536x865.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-500x282.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-1024x577.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Proteina-Complexa-Binder-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Proteina-Complexa-Binder-1920x1080" /><p>Developing new protein-based therapies and catalysts involves the challenging task of designing protein binders, or proteins that bind to a target protein or small molecule. The search space for possible amino acid sequence permutations and resulting 3D protein structures for a designed binder is vast, and achieving strong, specific binding requires careful optimization of the interactions between…</p>
<p><a href="https://developer.nvidia.com/blog/designing-protein-binders-using-the-generative-model-proteina-complexa/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/designing-protein-binders-using-the-generative-model-proteina-complexa/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/designing-protein-binders-using-the-generative-model-proteina-complexa/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Kibibi Moseley</name>
					</author>
		<title type="html"><![CDATA[Scaling Token Factory Revenue and AI Efficiency by Maximizing Performance per Watt]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/scaling-token-factory-revenue-and-ai-efficiency-by-maximizing-performance-per-watt/" />
		<id>https://developer.nvidia.com/blog/?p=114827</id>
		<updated>2026-04-16T17:15:27Z</updated>
		<published>2026-03-25T11:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Groq 3 LPX" /><category scheme="https://developer.nvidia.com/blog" term="Industrial Digitalization / Digital Twin" /><category scheme="https://developer.nvidia.com/blog" term="NVFP4" /><category scheme="https://developer.nvidia.com/blog" term="Sustainable Computing" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center" />In the AI era, power is the ultimate constraint, and every AI factory operates within a hard limit. This makes performance per watt—the rate at which power is...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/scaling-token-factory-revenue-and-ai-efficiency-by-maximizing-performance-per-watt/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center" />In the AI era, power is the ultimate constraint, and every AI factory operates within a hard limit. This makes performance per watt—the rate at which power is...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-2048x1152.jpg 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/data-center-960x540.jpg 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="data-center" /><p>In the AI era, power is the ultimate constraint, and every AI factory operates within a hard limit. This makes performance per watt—the rate at which power is converted into revenue-generating intelligence—the defining metric for modern AI infrastructure. AI data centers now operate as token factories tied directly to the energy ecosystem, where access to land, power…</p>
<p><a href="https://developer.nvidia.com/blog/scaling-token-factory-revenue-and-ai-efficiency-by-maximizing-performance-per-watt/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/scaling-token-factory-revenue-and-ai-efficiency-by-maximizing-performance-per-watt/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/scaling-token-factory-revenue-and-ai-efficiency-by-maximizing-performance-per-watt/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Chintan Patel</name>
					</author>
		<title type="html"><![CDATA[Building NVIDIA Nemotron 3 Agents for Reasoning, Multimodal RAG, Voice, and Safety]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-nvidia-nemotron-3-agents-for-reasoning-multimodal-rag-voice-and-safety/" />
		<id>https://developer.nvidia.com/blog/?p=114720</id>
		<updated>2026-04-16T17:15:29Z</updated>
		<published>2026-03-24T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Llama" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Machine Learning &amp; Artificial Intelligence" /><category scheme="https://developer.nvidia.com/blog" term="NeMo" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="NVFP4" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" /><category scheme="https://developer.nvidia.com/blog" term="Retrieval Augmented Generation (RAG)" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080 (1)" />Agentic AI is an ecosystem where specialized models work together to handle planning, reasoning, retrieval, and safety guardrailing. As these systems scale,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-nvidia-nemotron-3-agents-for-reasoning-multimodal-rag-voice-and-safety/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080 (1)" />Agentic AI is an ecosystem where specialized models work together to handle planning, reasoning, retrieval, and safety guardrailing. As these systems scale,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080 (1)" /><p>Agentic AI is an ecosystem where specialized models work together to handle planning, reasoning, retrieval, and safety guardrailing. As these systems scale, developers need models that can understand real-world multimodal data, converse naturally with users globally, and operate safely across languages and modalities. At GTC 2026, NVIDIA introduced a new generation of NVIDIA Nemotron models…</p>
<p><a href="https://developer.nvidia.com/blog/building-nvidia-nemotron-3-agents-for-reasoning-multimodal-rag-voice-and-safety/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-nvidia-nemotron-3-agents-for-reasoning-multimodal-rag-voice-and-safety/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-nvidia-nemotron-3-agents-for-reasoning-multimodal-rag-voice-and-safety/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Suhas Hariharapura Sheshadri</name>
						<uri>https://www.linkedin.com/in/suhassheshadri/</uri>
					</author>
		<title type="html"><![CDATA[NVIDIA IGX Thor Powers Industrial, Medical, and Robotics Edge AI Applications]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-igx-thor-powers-industrial-medical-and-robotics-edge-ai-applications/" />
		<id>https://developer.nvidia.com/blog/?p=112736</id>
		<updated>2026-04-16T17:15:31Z</updated>
		<published>2026-03-23T20:24:17Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Multi-Instance GPU (MIG)" /><category scheme="https://developer.nvidia.com/blog" term="Physical AI" /><category scheme="https://developer.nvidia.com/blog" term="Thor" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-igx-thor" />Industrial and medical systems are rapidly increasing the use of high-performance AI to improve worker productivity, human-machine interaction, and downtime...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-igx-thor-powers-industrial-medical-and-robotics-edge-ai-applications/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-igx-thor" />Industrial and medical systems are rapidly increasing the use of high-performance AI to improve worker productivity, human-machine interaction, and downtime...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-igx-thor.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-igx-thor" /><p>Industrial and medical systems are rapidly increasing the use of high-performance AI to improve worker productivity, human-machine interaction, and downtime management. From factory automation cells to autonomous mobile platforms to surgical rooms, operators are deploying increasingly complex generative AI models, more sensors, and higher‑fidelity data streams at the edge.</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-igx-thor-powers-industrial-medical-and-robotics-edge-ai-applications/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-igx-thor-powers-industrial-medical-and-robotics-edge-ai-applications/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-igx-thor-powers-industrial-medical-and-robotics-edge-ai-applications/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Hema Bontha</name>
					</author>
		<title type="html"><![CDATA[Building a Zero-Trust Architecture for Confidential AI Factories]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-a-zero-trust-architecture-for-confidential-ai-factories/" />
		<id>https://developer.nvidia.com/blog/?p=114591</id>
		<updated>2026-04-16T17:15:33Z</updated>
		<published>2026-03-23T12:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI / Cybersecurity" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="Confidential Compute" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cybersecurity-ai-featured" />AI is moving from experimentation to production. However, most data enterprises need exists outside the public cloud. This includes sensitive information like...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-a-zero-trust-architecture-for-confidential-ai-factories/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cybersecurity-ai-featured" />AI is moving from experimentation to production. However, most data enterprises need exists outside the public cloud. This includes sensitive information like...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/02/cybersecurity-ai-featured.png 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cybersecurity-ai-featured" /><p>AI is moving from experimentation to production. However, most data enterprises need exists outside the public cloud. This includes sensitive information like patient records, market research, and legacy systems containing enterprise knowledge. There’s also a risk of using private data with AI models, and adoption is often slowed or blocked by privacy and trust concerns.</p>
<p><a href="https://developer.nvidia.com/blog/building-a-zero-trust-architecture-for-confidential-ai-factories/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-a-zero-trust-architecture-for-confidential-ai-factories/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-a-zero-trust-architecture-for-confidential-ai-factories/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anish Maddipoti</name>
					</author>
		<title type="html"><![CDATA[Deploying Disaggregated LLM Inference Workloads on Kubernetes]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/deploying-disaggregated-llm-inference-workloads-on-kubernetes/" />
		<id>https://developer.nvidia.com/blog/?p=113609</id>
		<updated>2026-05-08T16:16:06Z</updated>
		<published>2026-03-23T07:01:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="AI Networking" /><category scheme="https://developer.nvidia.com/blog" term="Cloud Networking" /><category scheme="https://developer.nvidia.com/blog" term="Dynamo-Triton" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB200" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="Kubernetes" /><category scheme="https://developer.nvidia.com/blog" term="News" /><category scheme="https://developer.nvidia.com/blog" term="NVLink" /><category scheme="https://developer.nvidia.com/blog" term="Tutorial" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />As large language model (LLM) inference workloads grow in complexity, a single monolithic serving process starts to hit its limits. Prefill and decode stages...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/deploying-disaggregated-llm-inference-workloads-on-kubernetes/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" />As large language model (LLM) inference workloads grow in complexity, a single monolithic serving process starts to hit its limits. Prefill and decode stages...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image3-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image3" /><p>As large language model (LLM) inference workloads grow in complexity, a single monolithic serving process starts to hit its limits. Prefill and decode stages have fundamentally different compute profiles, yet traditional deployments force them onto the same hardware, leaving GPUs underutilized and scaling inflexible. Disaggregated serving addresses this by splitting the inference pipeline…</p>
<p><a href="https://developer.nvidia.com/blog/deploying-disaggregated-llm-inference-workloads-on-kubernetes/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/deploying-disaggregated-llm-inference-workloads-on-kubernetes/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/deploying-disaggregated-llm-inference-workloads-on-kubernetes/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Sean Lopp</name>
					</author>
		<title type="html"><![CDATA[How to Build Deep Agents for Enterprise Search with NVIDIA AI-Q and LangChain]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-deep-agents-for-enterprise-search-with-nvidia-ai-q-and-langchain/" />
		<id>https://developer.nvidia.com/blog/?p=114078</id>
		<updated>2026-04-16T17:15:35Z</updated>
		<published>2026-03-18T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Blueprint" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="NeMo" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="Retrieval Augmented Generation (RAG)" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080" />While consumer AI offers powerful capabilities, workplace tools often suffer from disjointed data and limited context. Built with LangChain, the NVIDIA AI-Q...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-build-deep-agents-for-enterprise-search-with-nvidia-ai-q-and-langchain/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080" />While consumer AI offers powerful capabilities, workplace tools often suffer from disjointed data and limited context. Built with LangChain, the NVIDIA AI-Q...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-aiq-blueprint-blog-gtc26-press-1920x1080" /><p>While consumer AI offers powerful capabilities, workplace tools often suffer from disjointed data and limited context. Built with LangChain, the NVIDIA AI-Q blueprint is an open source template that bridges this gap. LangChain recently introduced an enterprise agent platform built with NVIDIA AI to support scalable, production-ready agent development. This tutorial, available as an NVIDIA…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-build-deep-agents-for-enterprise-search-with-nvidia-ai-q-and-langchain/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-build-deep-agents-for-enterprise-search-with-nvidia-ai-q-and-langchain/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-build-deep-agents-for-enterprise-search-with-nvidia-ai-q-and-langchain/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Sree Sankar</name>
					</author>
		<title type="html"><![CDATA[Building the AI Grid with NVIDIA: Orchestrating Intelligence Everywhere ]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-the-ai-grid-with-nvidia-orchestrating-intelligence-everywhere/" />
		<id>https://developer.nvidia.com/blog/?p=114089</id>
		<updated>2026-04-16T17:15:37Z</updated>
		<published>2026-03-17T17:13:20Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="Cloud Networking" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Machine Learning &amp; Artificial Intelligence" /><category scheme="https://developer.nvidia.com/blog" term="Telecommunications" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="telco-promo-pack-ai-grid-tech-blog-1480x830" />AI-native services are exposing a new bottleneck in AI infrastructure: As millions of users, agents, and devices demand access to intelligence, the challenge is...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-the-ai-grid-with-nvidia-orchestrating-intelligence-everywhere/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="telco-promo-pack-ai-grid-tech-blog-1480x830" />AI-native services are exposing a new bottleneck in AI infrastructure: As millions of users, agents, and devices demand access to intelligence, the challenge is...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/telco-promo-pack-ai-grid-tech-blog-1480x830-1.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="telco-promo-pack-ai-grid-tech-blog-1480x830" /><p>AI-native services are exposing a new bottleneck in AI infrastructure: As millions of users, agents, and devices demand access to intelligence, the challenge is shifting from peak training throughput to delivering deterministic inference at scale—predictable latency, jitter, and sustainable token economics. NVIDIA announced at GTC 2026 that telcos and distributed cloud providers are…</p>
<p><a href="https://developer.nvidia.com/blog/building-the-ai-grid-with-nvidia-orchestrating-intelligence-everywhere/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-the-ai-grid-with-nvidia-orchestrating-intelligence-everywhere/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-the-ai-grid-with-nvidia-orchestrating-intelligence-everywhere/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Mingxin Zheng</name>
					</author>
		<title type="html"><![CDATA[Using Simulation to Build Robotic Systems for Hospital Automation]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/using-simulation-to-build-robotic-systems-for-hospital-automation/" />
		<id>https://developer.nvidia.com/blog/?p=114095</id>
		<updated>2026-04-16T17:15:39Z</updated>
		<published>2026-03-16T22:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GR00T" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Healthcare &amp; Life Sciences" /><category scheme="https://developer.nvidia.com/blog" term="Physical AI" /><category scheme="https://developer.nvidia.com/blog" term="Robotics Simulation" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="PeritasAI trains a DexMate Humanoid Robot at Advent Health hospital for sterilizing tools at a nursing station" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy.webp 1252w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Using Simulation to Build Robotic Systems for Hospital Automation" />Healthcare faces a structural demand–capacity crisis: a projected global shortfall of ~10 million clinicians by 2030, billions of diagnostic exams annually...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/using-simulation-to-build-robotic-systems-for-hospital-automation/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="PeritasAI trains a DexMate Humanoid Robot at Advent Health hospital for sterilizing tools at a nursing station" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy.webp 1252w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Using Simulation to Build Robotic Systems for Hospital Automation" />Healthcare faces a structural demand–capacity crisis: a projected global shortfall of ~10 million clinicians by 2030, billions of diagnostic exams annually...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="PeritasAI trains a DexMate Humanoid Robot at Advent Health hospital for sterilizing tools at a nursing station" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/image1-copy.webp 1252w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Using Simulation to Build Robotic Systems for Hospital Automation" /><p></p>
<p><a href="https://developer.nvidia.com/blog/using-simulation-to-build-robotic-systems-for-hospital-automation/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/using-simulation-to-build-robotic-systems-for-hospital-automation/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/using-simulation-to-build-robotic-systems-for-hospital-automation/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Amr Elmeleegy</name>
					</author>
		<title type="html"><![CDATA[How NVIDIA Dynamo 1.0 Powers Multi-Node Inference at Production Scale]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/" />
		<id>https://developer.nvidia.com/blog/?p=113961</id>
		<updated>2026-04-02T18:35:30Z</updated>
		<published>2026-03-16T20:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Agent toolkit" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="CUDA" /><category scheme="https://developer.nvidia.com/blog" term="Dynamo-Triton" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB200" /><category scheme="https://developer.nvidia.com/blog" term="GB300" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Hopper" /><category scheme="https://developer.nvidia.com/blog" term="Kubernetes" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="MLPerf" /><category scheme="https://developer.nvidia.com/blog" term="NVL72" /><category scheme="https://developer.nvidia.com/blog" term="NVLink" /><category scheme="https://developer.nvidia.com/blog" term="TensorRT-LLM" /><category scheme="https://developer.nvidia.com/blog" term="vLLM" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" />Reasoning models are growing rapidly in size and are increasingly being integrated into agentic AI workflows that interact with other models and external tools....]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" />Reasoning models are growing rapidly in size and are increasingly being integrated into agentic AI workflows that interact with other models and external tools....<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/inference-press-dynamo-gtc26-4960950-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="inference-press-dynamo-gtc26-4960950-1920x1080" /><p>Reasoning models are growing rapidly in size and are increasingly being integrated into agentic AI workflows that interact with other models and external tools. Deploying these models and workflows in production environments requires distributing them across multiple GPU nodes, which demands careful orchestration and coordination across GPUs. NVIDIA Dynamo 1.0—available now—addresses these…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Allen Bourgoyne</name>
					</author>
		<title type="html"><![CDATA[Scaling Autonomous AI Agents and Workloads with NVIDIA DGX Spark]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/scaling-autonomous-ai-agents-and-workloads-with-nvidia-dgx-spark/" />
		<id>https://developer.nvidia.com/blog/?p=114188</id>
		<updated>2026-04-16T17:15:40Z</updated>
		<published>2026-03-16T20:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="ConnectX" /><category scheme="https://developer.nvidia.com/blog" term="DGX Spark" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-2048x1152.png 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-960x540.png 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="four-stacked-nvidia-dgx-spark" />Autonomous AI agents are driving the next wave of AI innovation. These agents must often manage long-running tasks that use multiple communication channels and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/scaling-autonomous-ai-agents-and-workloads-with-nvidia-dgx-spark/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-2048x1152.png 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-960x540.png 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="four-stacked-nvidia-dgx-spark" />Autonomous AI agents are driving the next wave of AI innovation. These agents must often manage long-running tasks that use multiple communication channels and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-2048x1152.png 2048w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/four-stacked-nvidia-dgx-spark-960x540.png 960w" sizes="auto, (max-width: 768px) 100vw, 768px" title="four-stacked-nvidia-dgx-spark" /><p>Autonomous AI agents are driving the next wave of AI innovation. These agents must often manage long-running tasks that use multiple communication channels and background subprocesses simultaneously to explore options, test solutions, and generate optimal results. This places extreme demands on local compute. NVIDIA DGX Spark provides the performance necessary for autonomous agents to execute…</p>
<p><a href="https://developer.nvidia.com/blog/scaling-autonomous-ai-agents-and-workloads-with-nvidia-dgx-spark/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/scaling-autonomous-ai-agents-and-workloads-with-nvidia-dgx-spark/#comments" thr:count="2"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/scaling-autonomous-ai-agents-and-workloads-with-nvidia-dgx-spark/feed/" thr:count="2"/>
		<thr:total>2</thr:total>
	</entry>
		<entry>
		<author>
			<name>Moshe Anschel</name>
					</author>
		<title type="html"><![CDATA[Introducing NVIDIA BlueField-4-Powered CMX Context Memory Storage Platform for the Next Frontier of AI]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/introducing-nvidia-bluefield-4-powered-inference-context-memory-storage-platform-for-the-next-frontier-of-ai/" />
		<id>https://developer.nvidia.com/blog/?p=111143</id>
		<updated>2026-04-02T18:35:31Z</updated>
		<published>2026-03-16T20:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="CES26" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Rubin" /><category scheme="https://developer.nvidia.com/blog" term="Storage Networking &amp; Security" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-vera-rubin-nvl72" />AI‑native organizations increasingly face scaling challenges as agentic AI workflows drive context windows to millions of tokens and models scale toward...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/introducing-nvidia-bluefield-4-powered-inference-context-memory-storage-platform-for-the-next-frontier-of-ai/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-vera-rubin-nvl72" />AI‑native organizations increasingly face scaling challenges as agentic AI workflows drive context windows to millions of tokens and models scale toward...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/01/dgx-vera-rubin-nvl72.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="dgx-vera-rubin-nvl72" /><p>AI‑native organizations increasingly face scaling challenges as agentic AI workflows drive context windows to millions of tokens and models scale toward trillions of parameters. These systems rely on agentic long‑term memory for context that persists across turns, tools, and sessions so agents can build on prior reasoning instead of starting from scratch on every request.</p>
<p><a href="https://developer.nvidia.com/blog/introducing-nvidia-bluefield-4-powered-inference-context-memory-storage-platform-for-the-next-frontier-of-ai/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/introducing-nvidia-bluefield-4-powered-inference-context-memory-storage-platform-for-the-next-frontier-of-ai/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/introducing-nvidia-bluefield-4-powered-inference-context-memory-storage-platform-for-the-next-frontier-of-ai/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ranga Maddipudi</name>
					</author>
		<title type="html"><![CDATA[Design, Simulate, and Scale AI Factory Infrastructure with NVIDIA DSX Air]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/design-simulate-and-scale-ai-factory-infrastructure-with-nvidia-dsx-air/" />
		<id>https://developer.nvidia.com/blog/?p=113689</id>
		<updated>2026-04-02T18:35:32Z</updated>
		<published>2026-03-16T20:01:33Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Image of NVIDIA DSX Air being used on a laptop." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="DSX-Air" />Building AI factories is complex and requires efficient integration across compute, networking, security, and storage systems. To achieve rapid Time to AI and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/design-simulate-and-scale-ai-factory-infrastructure-with-nvidia-dsx-air/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Image of NVIDIA DSX Air being used on a laptop." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="DSX-Air" />Building AI factories is complex and requires efficient integration across compute, networking, security, and storage systems. To achieve rapid Time to AI and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Image of NVIDIA DSX Air being used on a laptop." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/DSX-Air.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="DSX-Air" /><p>Building AI factories is complex and requires efficient integration across compute, networking, security, and storage systems. To achieve rapid Time to AI and strong ROI, the new NVIDIA DSX Air is enabling organizations to simulate their entire AI factory infrastructure in the cloud—covering compute, networking, storage, and security. Being able to design, test, and optimize systems before…</p>
<p><a href="https://developer.nvidia.com/blog/design-simulate-and-scale-ai-factory-infrastructure-with-nvidia-dsx-air/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/design-simulate-and-scale-ai-factory-infrastructure-with-nvidia-dsx-air/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/design-simulate-and-scale-ai-factory-infrastructure-with-nvidia-dsx-air/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Praveen Menon</name>
					</author>
		<title type="html"><![CDATA[NVIDIA Vera CPU Delivers High Performance, Bandwidth, and Efficiency for AI Factories]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-vera-cpu-delivers-high-performance-bandwidth-and-efficiency-for-ai-factories/" />
		<id>https://developer.nvidia.com/blog/?p=114004</id>
		<updated>2026-04-21T15:53:20Z</updated>
		<published>2026-03-16T19:30:33Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Vera CPU" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Vera CPU render." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU.webp 2000w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Vera-CPU" />AI is evolving, and reasoning models are increasing token demand, placing new requirements on every layer of AI infrastructure. More than ever, compute must...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-vera-cpu-delivers-high-performance-bandwidth-and-efficiency-for-ai-factories/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Vera CPU render." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU.webp 2000w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Vera-CPU" />AI is evolving, and reasoning models are increasing token demand, placing new requirements on every layer of AI infrastructure. More than ever, compute must...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Vera CPU render." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Vera-CPU.webp 2000w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Vera-CPU" /><p>AI is evolving, and reasoning models are increasing token demand, placing new requirements on every layer of AI infrastructure. More than ever, compute must scale efficiently to maximize token production and improve productivity for model creators and users. Modern GPUs operate at peak capacity, pushing throughput higher every generation, but system performance is increasingly gated by the…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-vera-cpu-delivers-high-performance-bandwidth-and-efficiency-for-ai-factories/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-vera-cpu-delivers-high-performance-bandwidth-and-efficiency-for-ai-factories/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-vera-cpu-delivers-high-performance-bandwidth-and-efficiency-for-ai-factories/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ali Golshan</name>
					</author>
		<title type="html"><![CDATA[Run Autonomous, Self-Evolving Agents More Safely with NVIDIA OpenShell]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/" />
		<id>https://developer.nvidia.com/blog/?p=113924</id>
		<updated>2026-04-02T18:35:34Z</updated>
		<published>2026-03-16T16:10:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="Trustworthy AI / Cybersecurity" /><category scheme="https://developer.nvidia.com/blog" term="Agent toolkit" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="claws" /><category scheme="https://developer.nvidia.com/blog" term="DGX Spark" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="NeMo" /><category scheme="https://developer.nvidia.com/blog" term="NemoClaw" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" /><category scheme="https://developer.nvidia.com/blog" term="OpenShell" /><category scheme="https://developer.nvidia.com/blog" term="RTX GPU" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-enterprise-agents-gtc26-press-1920x1080" />AI has evolved from assistants following your directions to agents that act independently. Called claws, these agents can take a goal, figure out how to achieve...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-enterprise-agents-gtc26-press-1920x1080" />AI has evolved from assistants following your directions to agents that act independently. Called claws, these agents can take a goal, figure out how to achieve...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/agentic-ai-enterprise-agents-gtc26-press-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="agentic-ai-enterprise-agents-gtc26-press-1920x1080" /><p>AI has evolved from assistants following your directions to agents that act independently. Called claws, these agents can take a goal, figure out how to achieve it, and execute indefinitely—while leaving you out of the loop. The more capable claws become, the harder they are to trust. And their self-evolving autonomy changes everything about the environment in which they operate.</p>
<p><a href="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Kyle Aubrey</name>
					</author>
		<title type="html"><![CDATA[Inside NVIDIA Groq 3 LPX: The Low-Latency Inference Accelerator for the NVIDIA Vera Rubin Platform]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/" />
		<id>https://developer.nvidia.com/blog/?p=114202</id>
		<updated>2026-04-02T18:35:35Z</updated>
		<published>2026-03-16T16:09:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Groq 3 LPX" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Rubin GPU" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Render of LPX rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="LPX-Rack" />NVIDIA Groq 3 LPX is a new rack-scale inference accelerator for the NVIDIA Vera Rubin platform, designed for the low-latency and large-context demands of...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Render of LPX rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="LPX-Rack" />NVIDIA Groq 3 LPX is a new rack-scale inference accelerator for the NVIDIA Vera Rubin platform, designed for the low-latency and large-context demands of...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Render of LPX rack." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/LPX-Rack.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="LPX-Rack" /><p>NVIDIA Groq 3 LPX is a new rack-scale inference accelerator for the NVIDIA Vera Rubin platform, designed for the low-latency and large-context demands of agentic systems. Co-designed with the NVIDIA Vera Rubin NVL72, LPX equips the AI factory with an engine optimized for fast, predictable token generation, while Vera Rubin NVL72 remains the flexible, general-purpose workhorse for training and…</p>
<p><a href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Rohil Bhargava</name>
					</author>
		<title type="html"><![CDATA[NVIDIA Vera Rubin POD: Seven Chips, Five Rack-Scale Systems, One AI Supercomputer]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-vera-rubin-pod-seven-chips-five-rack-scale-systems-one-ai-supercomputer/" />
		<id>https://developer.nvidia.com/blog/?p=113993</id>
		<updated>2026-04-02T18:35:36Z</updated>
		<published>2026-03-16T16:05:58Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Factory" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="Rubin" /><category scheme="https://developer.nvidia.com/blog" term="Vera CPU" /><category scheme="https://developer.nvidia.com/blog" term="Vera Rubin NVL72" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" />Artificial intelligence is token-driven. Every prompt, reasoning step, and agent interaction generates tokens. Over the past year, token consumption has grown...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-vera-rubin-pod-seven-chips-five-rack-scale-systems-one-ai-supercomputer/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" />Artificial intelligence is token-driven. Every prompt, reasoning step, and agent interaction generates tokens. Over the past year, token consumption has grown...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/nvidia-vera-rubin-pod.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="nvidia-vera-rubin-pod" /><p>Artificial intelligence is token-driven. Every prompt, reasoning step, and agent interaction generates tokens. Over the past year, token consumption has grown multifold and now exceeds 10 quadrillion tokens per year. And while the majority of tokens have been generated from humans interacting with AI, the new era is one in which most tokens will be generated from AI interacting with AI.</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-vera-rubin-pod-seven-chips-five-rack-scale-systems-one-ai-supercomputer/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-vera-rubin-pod-seven-chips-five-rack-scale-systems-one-ai-supercomputer/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-vera-rubin-pod-seven-chips-five-rack-scale-systems-one-ai-supercomputer/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Philipp Reist</name>
					</author>
		<title type="html"><![CDATA[Newton Adds Contact-Rich Manipulation and Locomotion Capabilities for Industrial Robotics]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/newton-adds-contact-rich-manipulation-and-locomotion-capabilities-for-industrial-robotics/" />
		<id>https://developer.nvidia.com/blog/?p=113754</id>
		<updated>2026-04-02T18:35:37Z</updated>
		<published>2026-03-16T16:00:50Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp" class="webfeedsFeaturedVisual wp-post-image" alt="A Gif showing robot movements." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-195x110.webp 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Newton-Tasks" />Physics forms the foundation of robotic simulation, enabling realistic modeling of motion and interaction. For tasks like locomotion and manipulation,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/newton-adds-contact-rich-manipulation-and-locomotion-capabilities-for-industrial-robotics/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp" class="webfeedsFeaturedVisual wp-post-image" alt="A Gif showing robot movements." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-195x110.webp 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Newton-Tasks" />Physics forms the foundation of robotic simulation, enabling realistic modeling of motion and interaction. For tasks like locomotion and manipulation,...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp" class="webfeedsFeaturedVisual wp-post-image" alt="A Gif showing robot movements." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks.webp 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Newton-Tasks-195x110.webp 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="Newton-Tasks" /><p>Physics forms the foundation of robotic simulation, enabling realistic modeling of motion and interaction. For tasks like locomotion and manipulation, simulators must handle complex dynamics such as contact forces and deformable objects. While most engines trade off speed for realism, Newton—a GPU-accelerated, open source simulator—is designed to do both. Newton 1.0 GA…</p>
<p><a href="https://developer.nvidia.com/blog/newton-adds-contact-rich-manipulation-and-locomotion-capabilities-for-industrial-robotics/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
<link href="https://developer.download.nvidia.com/video/devblog/Tactile.mp4" rel="enclosure" length="2597996" type="video/mp4" />
<link href="https://developer.download.nvidia.com/video/devblog/Warp.mp4" rel="enclosure" length="14542963" type="video/mp4" />
<link href="https://developer.download.nvidia.com/video/devblog/kamino-dr-legs.mp4" rel="enclosure" length="16339398" type="video/mp4" />
<link href="https://developer.download.nvidia.com/video/devblog/skild-busbar.mp4" rel="enclosure" length="109763230" type="video/mp4" />
<link href="https://developer.download.nvidia.com/video/devblog/waterhose.mp4" rel="enclosure" length="20213115" type="video/mp4" />
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/newton-adds-contact-rich-manipulation-and-locomotion-capabilities-for-industrial-robotics/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/newton-adds-contact-rich-manipulation-and-locomotion-capabilities-for-industrial-robotics/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Pranjali Joshi</name>
					</author>
		<title type="html"><![CDATA[Scale Synthetic Data and Physical AI Reasoning with NVIDIA Cosmos World Foundation Models]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/scale-synthetic-data-and-physical-ai-reasoning-with-nvidia-cosmos-world-foundation-models/" />
		<id>https://developer.nvidia.com/blog/?p=97132</id>
		<updated>2026-04-02T18:35:38Z</updated>
		<published>2026-03-13T16:00:47Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Cosmos" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/03/Cosmos-Data-Reasoning.gif" class="webfeedsFeaturedVisual wp-post-image" alt="A GIF showing robotics." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="Cosmos-Data-Reasoning" />The next generation of AI-driven robots like humanoids and autonomous vehicles depends on high-fidelity, physics-aware training data. Without diverse and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/scale-synthetic-data-and-physical-ai-reasoning-with-nvidia-cosmos-world-foundation-models/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/03/Cosmos-Data-Reasoning.gif" class="webfeedsFeaturedVisual wp-post-image" alt="A GIF showing robotics." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="Cosmos-Data-Reasoning" />The next generation of AI-driven robots like humanoids and autonomous vehicles depends on high-fidelity, physics-aware training data. Without diverse and...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/03/Cosmos-Data-Reasoning.gif" class="webfeedsFeaturedVisual wp-post-image" alt="A GIF showing robotics." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="Cosmos-Data-Reasoning" /><p>The next generation of AI-driven robots like humanoids and autonomous vehicles depends on high-fidelity, physics-aware training data. Without diverse and representative datasets, these systems don’t get proper training and face testing risks due to poor generalization, limited exposure to real-world variations, and unpredictable behavior in edge cases. Collecting massive real-world datasets for…</p>
<p><a href="https://developer.nvidia.com/blog/scale-synthetic-data-and-physical-ai-reasoning-with-nvidia-cosmos-world-foundation-models/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/scale-synthetic-data-and-physical-ai-reasoning-with-nvidia-cosmos-world-foundation-models/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/scale-synthetic-data-and-physical-ai-reasoning-with-nvidia-cosmos-world-foundation-models/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Sheel Nidhan</name>
					</author>
		<title type="html"><![CDATA[Build Accelerated, Differentiable Computational Physics Code for AI with NVIDIA Warp]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-accelerated-differentiable-computational-physics-code-for-ai-with-nvidia-warp/" />
		<id>https://developer.nvidia.com/blog/?p=113051</id>
		<updated>2026-04-02T18:35:39Z</updated>
		<published>2026-03-12T17:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="CAE" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Physics" /><category scheme="https://developer.nvidia.com/blog" term="Python" /><category scheme="https://developer.nvidia.com/blog" term="research" /><category scheme="https://developer.nvidia.com/blog" term="Warp" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="decaying-turbulence" />Computer-aided engineering (CAE) is shifting from human-driven workflows toward AI-driven ones, including physics foundation models that generalize across...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-accelerated-differentiable-computational-physics-code-for-ai-with-nvidia-warp/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="decaying-turbulence" />Computer-aided engineering (CAE) is shifting from human-driven workflows toward AI-driven ones, including physics foundation models that generalize across...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/decaying-turbulence.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="decaying-turbulence" /><p>Computer-aided engineering (CAE) is shifting from human-driven workflows toward AI-driven ones, including physics foundation models that generalize across geometries and operating conditions. Unlike LLMs, these models depend on large volumes of high-fidelity, physics-compliant data. Recent scaling-law work on computational fluid dynamics (CFD) surrogates indicates that simulation-generated…</p>
<p><a href="https://developer.nvidia.com/blog/build-accelerated-differentiable-computational-physics-code-for-ai-with-nvidia-warp/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-accelerated-differentiable-computational-physics-code-for-ai-with-nvidia-warp/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-accelerated-differentiable-computational-physics-code-for-ai-with-nvidia-warp/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Mark Chmarny</name>
					</author>
		<title type="html"><![CDATA[Validate Kubernetes for GPU Infrastructure with Layered, Reproducible Recipes]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/validate-kubernetes-for-gpu-infrastructure-with-layered-reproducible-recipes/" />
		<id>https://developer.nvidia.com/blog/?p=112987</id>
		<updated>2026-04-02T18:35:40Z</updated>
		<published>2026-03-12T16:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="Cloud Services" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Kubernetes" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-1024x575.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405.webp 1054w" sizes="auto, (max-width: 768px) 100vw, 768px" title="security-social-cve-workflow-3882621-1200x628" />Every AI cluster running on Kubernetes requires a full software stack that works together, from low-level driver and kernel settings to high-level operator and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/validate-kubernetes-for-gpu-infrastructure-with-layered-reproducible-recipes/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-1024x575.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405.webp 1054w" sizes="auto, (max-width: 768px) 100vw, 768px" title="security-social-cve-workflow-3882621-1200x628" />Every AI cluster running on Kubernetes requires a full software stack that works together, from low-level driver and kernel settings to high-level operator and...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-768x431.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-645x362.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-660x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-362x203.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-1024x575.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/security-social-cve-workflow-3882621-1200x628-1-e1772213795405.webp 1054w" sizes="auto, (max-width: 768px) 100vw, 768px" title="security-social-cve-workflow-3882621-1200x628" /><p>Every AI cluster running on Kubernetes requires a full software stack that works together, from low-level driver and kernel settings to high-level operator and workload configurations. You get one cluster working, and spend days getting the next one to match. Upgrade a component, and something else breaks. Move to a new cloud and start over. AI Cluster Runtime is a new open-source project designed…</p>
<p><a href="https://developer.nvidia.com/blog/validate-kubernetes-for-gpu-infrastructure-with-layered-reproducible-recipes/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/validate-kubernetes-for-gpu-infrastructure-with-layered-reproducible-recipes/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/validate-kubernetes-for-gpu-infrastructure-with-layered-reproducible-recipes/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Lin Chai</name>
					</author>
		<title type="html"><![CDATA[Build Next-Gen Physical AI with Edge‑First LLMs for Autonomous Vehicles and Robotics]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-next-gen-physical-ai-with-edge%e2%80%91first-llms-for-autonomous-vehicles-and-robotics/" />
		<id>https://developer.nvidia.com/blog/?p=113671</id>
		<updated>2026-03-12T00:11:11Z</updated>
		<published>2026-03-12T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="autonomous vehicles" /><category scheme="https://developer.nvidia.com/blog" term="GTC 2026" /><category scheme="https://developer.nvidia.com/blog" term="IoT" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Mixture of Experts (MoE)" /><category scheme="https://developer.nvidia.com/blog" term="Physical AI" /><category scheme="https://developer.nvidia.com/blog" term="Retrieval Augmented Generation (RAG)" /><category scheme="https://developer.nvidia.com/blog" term="Thor" /><category scheme="https://developer.nvidia.com/blog" term="VLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="autonomous-vehicle-backseat-view" />Physical AI is rapidly evolving, from next-generation software-defined autonomous vehicles (AVs) to humanoid robots. The challenge is no longer how to run a...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-next-gen-physical-ai-with-edge%e2%80%91first-llms-for-autonomous-vehicles-and-robotics/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="autonomous-vehicle-backseat-view" />Physical AI is rapidly evolving, from next-generation software-defined autonomous vehicles (AVs) to humanoid robots. The challenge is no longer how to run a...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/autonomous-vehicle-backseat-view.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="autonomous-vehicle-backseat-view" /><p>Physical AI is rapidly evolving, from next-generation software-defined autonomous vehicles (AVs) to humanoid robots. The challenge is no longer how to run a large language model (LLM), but how to enable high-fidelity reasoning, real-time multimodal interaction, and trajectory planning within strict power and latency envelopes. NVIDIA TensorRT Edge-LLM, a high-performance C++ inference runtime…</p>
<p><a href="https://developer.nvidia.com/blog/build-next-gen-physical-ai-with-edge%e2%80%91first-llms-for-autonomous-vehicles-and-robotics/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-next-gen-physical-ai-with-edge%e2%80%91first-llms-for-autonomous-vehicles-and-robotics/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-next-gen-physical-ai-with-edge%e2%80%91first-llms-for-autonomous-vehicles-and-robotics/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Chris Alexiuk</name>
					</author>
		<title type="html"><![CDATA[Introducing Nemotron 3 Super: An Open Hybrid Mamba-Transformer MoE for Agentic Reasoning]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/" />
		<id>https://developer.nvidia.com/blog/?p=113379</id>
		<updated>2026-03-12T17:36:42Z</updated>
		<published>2026-03-11T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLM Benchmarking" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Machine Learning &amp; Artificial Intelligence" /><category scheme="https://developer.nvidia.com/blog" term="Mixture of Experts (MoE)" /><category scheme="https://developer.nvidia.com/blog" term="NeMo" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="News" /><category scheme="https://developer.nvidia.com/blog" term="NIM" /><category scheme="https://developer.nvidia.com/blog" term="NVFP4" /><category scheme="https://developer.nvidia.com/blog" term="Reinforcement Learning" /><category scheme="https://developer.nvidia.com/blog" term="TensorRT-LLM" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080" />Agentic AI systems need models with the specialized depth to solve dense technical problems autonomously. They must excel at reasoning, coding, and long-context...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080" />Agentic AI systems need models with the specialized depth to solve dense technical problems autonomously. They must excel at reasoning, coding, and long-context...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Copy-of-genai-social-nemotron-3-4643900-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Copy of genai-social-nemotron-3-4643900-1920x1080" /><p>Agentic AI systems need models with the specialized depth to solve dense technical problems autonomously. They must excel at reasoning, coding, and long-context analysis, while remaining efficient enough to run continuously at scale. Multi-agent systems generate up to 15x the tokens of standard chats, re-sending history, tool outputs, and reasoning steps at every turn. Over long tasks…</p>
<p><a href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Ike Nnoli</name>
					</author>
		<title type="html"><![CDATA[NVIDIA RTX Innovations Are Powering the Next Era of Game Development]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/nvidia-rtx-innovations-are-powering-the-next-era-of-game-development/" />
		<id>https://developer.nvidia.com/blog/?p=113506</id>
		<updated>2026-03-12T20:47:03Z</updated>
		<published>2026-03-10T15:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="DirectX" /><category scheme="https://developer.nvidia.com/blog" term="GDC" /><category scheme="https://developer.nvidia.com/blog" term="GeForce" /><category scheme="https://developer.nvidia.com/blog" term="Neural Graphics" /><category scheme="https://developer.nvidia.com/blog" term="NvRTX" /><category scheme="https://developer.nvidia.com/blog" term="Ray Tracing / Path Tracing" /><category scheme="https://developer.nvidia.com/blog" term="Text Processing" /><category scheme="https://developer.nvidia.com/blog" term="Unreal Engine" /><category scheme="https://developer.nvidia.com/blog" term="vGPU" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="foliage-mountain-gif" />NVIDIA RTX ray tracing and AI-powered neural rendering technologies are redefining how games are made, enabling a new standard for visuals and performance. At...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/nvidia-rtx-innovations-are-powering-the-next-era-of-game-development/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="foliage-mountain-gif" />NVIDIA RTX ray tracing and AI-powered neural rendering technologies are redefining how games are made, enabling a new standard for visuals and performance. At...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/foliage-mountain-gif-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="foliage-mountain-gif" /><p>NVIDIA RTX ray tracing and AI-powered neural rendering technologies are redefining how games are made, enabling a new standard for visuals and performance. At GDC 2026, NVIDIA unveiled the latest path tracing innovations elevating visual fidelity, on-device AI models enabling players to interact with their favorite experiences in new ways, and enterprise solutions accelerating game development…</p>
<p><a href="https://developer.nvidia.com/blog/nvidia-rtx-innovations-are-powering-the-next-era-of-game-development/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/nvidia-rtx-innovations-are-powering-the-next-era-of-game-development/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/nvidia-rtx-innovations-are-powering-the-next-era-of-game-development/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Paul Logan</name>
					</author>
		<title type="html"><![CDATA[Reliable AI Coding for Unreal Engine: Improving Accuracy and Reducing Token Costs]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/reliable-ai-coding-for-unreal-engine-improving-accuracy-and-reducing-token-costs/" />
		<id>https://developer.nvidia.com/blog/?p=113547</id>
		<updated>2026-03-09T19:47:04Z</updated>
		<published>2026-03-10T15:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Content Creation / Rendering" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="C++" /><category scheme="https://developer.nvidia.com/blog" term="Unreal Engine" />		<summary type="html"><![CDATA[<img width="768" height="433" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-1024x577.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Reliable-AI-Coding" />Agentic code assistants are moving into daily game development as studios build larger worlds, ship more DLCs, and support distributed teams. These assistants...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/reliable-ai-coding-for-unreal-engine-improving-accuracy-and-reducing-token-costs/"><![CDATA[<img width="768" height="433" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-1024x577.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Reliable-AI-Coding" />Agentic code assistants are moving into daily game development as studios build larger worlds, ship more DLCs, and support distributed teams. These assistants...<img width="768" height="433" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-768x433.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-625x352.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-500x282.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-195x110.webp 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-1024x577.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Reliable-AI-Coding-e1772828712460.webp 1536w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Reliable-AI-Coding" /><p>Agentic code assistants are moving into daily game development as studios build larger worlds, ship more DLCs, and support distributed teams. These assistants can accelerate development by helping with tasks like generating gameplay scaffolding, refactoring repetitive systems, and answering engine-specific questions faster. This post outlines how developers can build reliable AI coding…</p>
<p><a href="https://developer.nvidia.com/blog/reliable-ai-coding-for-unreal-engine-improving-accuracy-and-reducing-token-costs/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/reliable-ai-coding-for-unreal-engine-improving-accuracy-and-reducing-token-costs/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/reliable-ai-coding-for-unreal-engine-improving-accuracy-and-reducing-token-costs/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Jonathan Bentz</name>
					</author>
		<title type="html"><![CDATA[CUDA 13.2 Introduces Enhanced CUDA Tile Support and New Python Features]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/cuda-13-2-introduces-enhanced-cuda-tile-support-and-new-python-features/" />
		<id>https://developer.nvidia.com/blog/?p=112653</id>
		<updated>2026-03-09T21:13:43Z</updated>
		<published>2026-03-09T21:13:18Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="CUDA C++" /><category scheme="https://developer.nvidia.com/blog" term="CUDA Tile" /><category scheme="https://developer.nvidia.com/blog" term="Memory" /><category scheme="https://developer.nvidia.com/blog" term="Multi-Instance GPU (MIG)" /><category scheme="https://developer.nvidia.com/blog" term="Python" /><category scheme="https://developer.nvidia.com/blog" term="Release" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube" />CUDA 13.2 arrives with a major update: NVIDIA CUDA Tile is now supported on devices of compute capability 8.X architectures (NVIDIA Ampere and NVIDIA Ada), as...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/cuda-13-2-introduces-enhanced-cuda-tile-support-and-new-python-features/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube" />CUDA 13.2 arrives with a major update: NVIDIA CUDA Tile is now supported on devices of compute capability 8.X architectures (NVIDIA Ampere and NVIDIA Ada), as...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cube-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cube" /><p>CUDA 13.2 arrives with a major update: NVIDIA CUDA Tile is now supported on devices of compute capability 8.X architectures (NVIDIA Ampere and NVIDIA Ada), as well as 10.X, 11.X and 12.X architectures (NVIDIA Blackwell). In an upcoming release of the CUDA Toolkit, all GPU architectures starting with Ampere will be fully supported. If you’re using Ampere, Ada, or Blackwell GPU architectures…</p>
<p><a href="https://developer.nvidia.com/blog/cuda-13-2-introduces-enhanced-cuda-tile-support-and-new-python-features/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/cuda-13-2-introduces-enhanced-cuda-tile-support-and-new-python-features/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/cuda-13-2-introduces-enhanced-cuda-tile-support-and-new-python-features/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Mireille Fares</name>
					</author>
		<title type="html"><![CDATA[Implementing Falcon-H1 Hybrid Architecture in NVIDIA Megatron Core]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/implementing-falcon-h1-hybrid-architecture-in-nvidia-megatron-core/" />
		<id>https://developer.nvidia.com/blog/?p=113474</id>
		<updated>2026-03-09T18:07:24Z</updated>
		<published>2026-03-09T19:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="AI Foundation Models" /><category scheme="https://developer.nvidia.com/blog" term="Megatron" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." />In the rapidly evolving landscape of large language model (LLM) development, NVIDIA Megatron Core has emerged as the foundational framework for training massive...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/implementing-falcon-h1-hybrid-architecture-in-nvidia-megatron-core/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." />In the rapidly evolving landscape of large language model (LLM) development, NVIDIA Megatron Core has emerged as the foundational framework for training massive...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2024/07/stacked-geometric-shapes-1.jpg 1209w" sizes="auto, (max-width: 768px) 100vw, 768px" title="stacked-geometric-shapes." /><p>In the rapidly evolving landscape of large language model (LLM) development, NVIDIA Megatron Core has emerged as the foundational framework for training massive transformer models at scale. The open source library offers industry-leading parallelism and GPU-optimized performance. Now developed GitHub-first in the NVIDIA/Megatron-LM repo, Megatron Core is increasingly shaped by contributions from…</p>
<p><a href="https://developer.nvidia.com/blog/implementing-falcon-h1-hybrid-architecture-in-nvidia-megatron-core/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/implementing-falcon-h1-hybrid-architecture-in-nvidia-megatron-core/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/implementing-falcon-h1-hybrid-architecture-in-nvidia-megatron-core/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Seonghee Lee</name>
					</author>
		<title type="html"><![CDATA[Enhancing Distributed Inference Performance with the NVIDIA Inference Transfer Library]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/enhancing-distributed-inference-performance-with-the-nvidia-inference-transfer-library/" />
		<id>https://developer.nvidia.com/blog/?p=113426</id>
		<updated>2026-03-06T18:29:04Z</updated>
		<published>2026-03-09T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="Python" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-data" />Deploying large language models (LLMs) requires large-scale distributed inference, which spreads model computation and request handling across many GPUs and...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/enhancing-distributed-inference-performance-with-the-nvidia-inference-transfer-library/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-data" />Deploying large language models (LLMs) requires large-scale distributed inference, which spreads model computation and request handling across many GPUs and...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/ai-data.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-data" /><p>Deploying large language models (LLMs) requires large-scale distributed inference, which spreads model computation and request handling across many GPUs and nodes to scale to more users while reducing latency. Distributed inference frameworks use techniques such as disaggregated serving, KV cache loading, and wide expert parallelism. In disaggregated serving environments…</p>
<p><a href="https://developer.nvidia.com/blog/enhancing-distributed-inference-performance-with-the-nvidia-inference-transfer-library/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/enhancing-distributed-inference-performance-with-the-nvidia-inference-transfer-library/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/enhancing-distributed-inference-performance-with-the-nvidia-inference-transfer-library/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Tianhao Xu</name>
					</author>
		<title type="html"><![CDATA[Removing the Guesswork from Disaggregated Serving]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/removing-the-guesswork-from-disaggregated-serving/" />
		<id>https://developer.nvidia.com/blog/?p=113333</id>
		<updated>2026-03-09T15:13:35Z</updated>
		<published>2026-03-09T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="A100" /><category scheme="https://developer.nvidia.com/blog" term="Dynamo" /><category scheme="https://developer.nvidia.com/blog" term="GB200" /><category scheme="https://developer.nvidia.com/blog" term="GB300" /><category scheme="https://developer.nvidia.com/blog" term="H100" /><category scheme="https://developer.nvidia.com/blog" term="LLM Techniques" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Machine Learning &amp; Artificial Intelligence" /><category scheme="https://developer.nvidia.com/blog" term="Mixture of Experts (MoE)" /><category scheme="https://developer.nvidia.com/blog" term="TensorRT-LLM" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-mixture-of-experts-blog-3105601-1920x1080" />Deploying and optimizing large language models (LLMs) for high-performance, cost-effective serving can be an overwhelming engineering problem. The ideal...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/removing-the-guesswork-from-disaggregated-serving/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-mixture-of-experts-blog-3105601-1920x1080" />Deploying and optimizing large language models (LLMs) for high-performance, cost-effective serving can be an overwhelming engineering problem. The ideal...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/genai-mixture-of-experts-blog-3105601-1920x1080-1.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-mixture-of-experts-blog-3105601-1920x1080" /><p>Deploying and optimizing large language models (LLMs) for high-performance, cost-effective serving can be an overwhelming engineering problem. The ideal configuration for any given workload (such as hardware, parallelism, and prefill/decode split) resides in a massive, multi-dimensional search space that is impossible to explore manually or through exhaustive testing. AIConfigurator…</p>
<p><a href="https://developer.nvidia.com/blog/removing-the-guesswork-from-disaggregated-serving/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/removing-the-guesswork-from-disaggregated-serving/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/removing-the-guesswork-from-disaggregated-serving/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Alessandro Morari</name>
					</author>
		<title type="html"><![CDATA[Tuning Flash Attention for Peak Performance in NVIDIA CUDA Tile]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/tuning-flash-attention-for-peak-performance-in-nvidia-cuda-tile/" />
		<id>https://developer.nvidia.com/blog/?p=113179</id>
		<updated>2026-03-05T19:48:13Z</updated>
		<published>2026-03-05T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="CUDA Tile" /><category scheme="https://developer.nvidia.com/blog" term="cuTile" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1536x862.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1024x575.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention.webp 1837w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA-Tile-Flash-Attention" />In this post, we dive into one of the most critical workloads in modern AI: Flash Attention, where you’ll learn: How to implement Flash Attention using NVIDIA...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/tuning-flash-attention-for-peak-performance-in-nvidia-cuda-tile/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1536x862.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1024x575.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention.webp 1837w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA-Tile-Flash-Attention" />In this post, we dive into one of the most critical workloads in modern AI: Flash Attention, where you’ll learn: How to implement Flash Attention using NVIDIA...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1536x862.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-1024x575.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/CUDA-Tile-Flash-Attention.webp 1837w" sizes="auto, (max-width: 768px) 100vw, 768px" title="CUDA-Tile-Flash-Attention" /><p>In this post, we dive into one of the most critical workloads in modern AI: Flash Attention, where you’ll learn: Environment requirements: See the quickstart doc for more information on installing cuTile Python. The attention mechanism is the computational heart of transformer models. Given a sequence of tokens, attention enables each token to “look at” every other…</p>
<p><a href="https://developer.nvidia.com/blog/tuning-flash-attention-for-peak-performance-in-nvidia-cuda-tile/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/tuning-flash-attention-for-peak-performance-in-nvidia-cuda-tile/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/tuning-flash-attention-for-peak-performance-in-nvidia-cuda-tile/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Nader Al Awar</name>
					</author>
		<title type="html"><![CDATA[Controlling Floating-Point Determinism in NVIDIA CCCL]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/controlling-floating-point-determinism-in-nvidia-cccl/" />
		<id>https://developer.nvidia.com/blog/?p=113316</id>
		<updated>2026-03-05T19:19:41Z</updated>
		<published>2026-03-05T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Floating-Point-CUB" />A computation is considered deterministic if multiple runs with the same input data produce the same bitwise result. While this may seem like a simple property...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/controlling-floating-point-determinism-in-nvidia-cccl/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Floating-Point-CUB" />A computation is considered deterministic if multiple runs with the same input data produce the same bitwise result. While this may seem like a simple property...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Floating-Point-CUB.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Floating-Point-CUB" /><p></p>
<p><a href="https://developer.nvidia.com/blog/controlling-floating-point-determinism-in-nvidia-cccl/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/controlling-floating-point-determinism-in-nvidia-cccl/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/controlling-floating-point-determinism-in-nvidia-cccl/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Brandon Rowlett</name>
					</author>
		<title type="html"><![CDATA[How to Minimize Game Runtime Inference Costs with Coding Agents]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-to-minimize-game-runtime-inference-costs-with-coding-agents/" />
		<id>https://developer.nvidia.com/blog/?p=113159</id>
		<updated>2026-03-05T19:19:43Z</updated>
		<published>2026-03-03T19:49:57Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="DLSS" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Game Performance" /><category scheme="https://developer.nvidia.com/blog" term="RTX AI" /><category scheme="https://developer.nvidia.com/blog" term="SLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative-image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Inference-Game-Agents" />NVIDIA ACE is a suite of technologies for building AI agents for gaming. ACE provides ready-to-integrate cloud and on-device AI models for every part of in-game...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-to-minimize-game-runtime-inference-costs-with-coding-agents/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative-image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Inference-Game-Agents" />NVIDIA ACE is a suite of technologies for building AI agents for gaming. ACE provides ready-to-integrate cloud and on-device AI models for every part of in-game...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="Decorative-image." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/03/Inference-Game-Agents.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="Inference-Game-Agents" /><p>NVIDIA ACE is a suite of technologies for building AI agents for gaming. ACE provides ready-to-integrate cloud and on-device AI models for every part of in-game characters, from speech to intelligence to animation. To run these models alongside the game engine efficiently, the NVIDIA In-Game Inferencing (NVIGI) SDK includes a set of performant libraries that developers can integrate into C++…</p>
<p><a href="https://developer.nvidia.com/blog/how-to-minimize-game-runtime-inference-costs-with-coding-agents/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-to-minimize-game-runtime-inference-costs-with-coding-agents/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-to-minimize-game-runtime-inference-costs-with-coding-agents/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Tim Besard</name>
					</author>
		<title type="html"><![CDATA[cuTile.jl Brings NVIDIA CUDA Tile-Based Programming to Julia]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/cutile-jl-brings-nvidia-cuda-tile-based-programming-to-julia/" />
		<id>https://developer.nvidia.com/blog/?p=112860</id>
		<updated>2026-03-05T19:45:49Z</updated>
		<published>2026-03-03T19:48:16Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="CUDA Tile" /><category scheme="https://developer.nvidia.com/blog" term="cuTile" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Python" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="JuliaHub logo." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub.webp 1514w" sizes="auto, (max-width: 768px) 100vw, 768px" title="JuliaHub" />NVIDIA CUDA Tile is one of the most significant additions to NVIDIA CUDA programming and unlocks automatic access to tensor cores and other specialized...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/cutile-jl-brings-nvidia-cuda-tile-based-programming-to-julia/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="JuliaHub logo." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub.webp 1514w" sizes="auto, (max-width: 768px) 100vw, 768px" title="JuliaHub" />NVIDIA CUDA Tile is one of the most significant additions to NVIDIA CUDA programming and unlocks automatic access to tensor cores and other specialized...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="JuliaHub logo." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/JuliaHub.webp 1514w" sizes="auto, (max-width: 768px) 100vw, 768px" title="JuliaHub" /><p>NVIDIA CUDA Tile is one of the most significant additions to NVIDIA CUDA programming and unlocks automatic access to tensor cores and other specialized hardware. Earlier this year, NVIDIA released cuTile for Python, giving Python developers a natural way to write high-performance GPU kernels. Now, the same programming model is available in Julia through cuTile.jl. In this blog post…</p>
<p><a href="https://developer.nvidia.com/blog/cutile-jl-brings-nvidia-cuda-tile-based-programming-to-julia/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/cutile-jl-brings-nvidia-cuda-tile-based-programming-to-julia/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/cutile-jl-brings-nvidia-cuda-tile-based-programming-to-julia/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Aiden Chang</name>
					</author>
		<title type="html"><![CDATA[Building Telco Reasoning Models for Autonomous Networks with NVIDIA NeMo]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/building-telco-reasoning-models-for-autonomous-networks-with-nvidia-nemo/" />
		<id>https://developer.nvidia.com/blog/?p=112936</id>
		<updated>2026-03-05T19:19:46Z</updated>
		<published>2026-03-01T07:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Retrieval Augmented Generation (RAG)" /><category scheme="https://developer.nvidia.com/blog" term="Training AI Models" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-reasoning-networking" />Autonomous networks are quickly becoming one of the top priorities in telecommunications. According to the latest NVIDIA State of AI in Telecommunications...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/building-telco-reasoning-models-for-autonomous-networks-with-nvidia-nemo/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-reasoning-networking" />Autonomous networks are quickly becoming one of the top priorities in telecommunications. According to the latest NVIDIA State of AI in Telecommunications...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ai-reasoning-networking.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ai-reasoning-networking" /><p>Autonomous networks are quickly becoming one of the top priorities in telecommunications. According to the latest NVIDIA State of AI in Telecommunications report, 65% of operators said AI is driving network automation, and 50% named autonomous networks as the top AI use case for ROI. Yet many telcos still report gaps in AI and data science expertise. This makes it difficult to scale safe…</p>
<p><a href="https://developer.nvidia.com/blog/building-telco-reasoning-models-for-autonomous-networks-with-nvidia-nemo/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/building-telco-reasoning-models-for-autonomous-networks-with-nvidia-nemo/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/building-telco-reasoning-models-for-autonomous-networks-with-nvidia-nemo/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Cindy Goh</name>
					</author>
		<title type="html"><![CDATA[5 New Digital Twin Products Developers Can Use to Build 6G Networks]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/5-new-digital-twin-products-developers-can-use-to-build-6g-networks/" />
		<id>https://developer.nvidia.com/blog/?p=113018</id>
		<updated>2026-03-25T23:15:33Z</updated>
		<published>2026-03-01T07:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Networking / Communications" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="5G / 6G" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Industrial Digitalization / Digital Twin" />		<summary type="html"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="A 3D visualization of a digital twin of a city." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image6" />To make 6G a reality, the telecom industry must overcome a fundamental challenge: how to design, train, and validate AI-native networks that are too complex to...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/5-new-digital-twin-products-developers-can-use-to-build-6g-networks/"><![CDATA[<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="A 3D visualization of a digital twin of a city." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image6" />To make 6G a reality, the telecom industry must overcome a fundamental challenge: how to design, train, and validate AI-native networks that are too complex to...<img width="768" height="431" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png" class="webfeedsFeaturedVisual wp-post-image" alt="A 3D visualization of a digital twin of a city." style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-768x431.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-179x100.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-300x168.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-625x351.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-645x362.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-500x280.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-362x203.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-1024x574.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6-960x538.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image6.webp 1480w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image6" /><p>To make 6G a reality, the telecom industry must overcome a fundamental challenge: how to design, train, and validate AI-native networks that are too complex to be tested in the physical world. The NVIDIA Aerial Omniverse Digital Twin (AODT) solves this by enabling a continuous integration/continuous development (CI/CD)-style workflow where Radio Access Network (RAN) software is trained…</p>
<p><a href="https://developer.nvidia.com/blog/5-new-digital-twin-products-developers-can-use-to-build-6g-networks/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/5-new-digital-twin-products-developers-can-use-to-build-6g-networks/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/5-new-digital-twin-products-developers-can-use-to-build-6g-networks/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Anu Srivastava</name>
					</author>
		<title type="html"><![CDATA[Develop Native Multimodal Agents with Qwen3.5 VLM Using NVIDIA GPU-Accelerated Endpoints]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/develop-native-multimodal-agents-with-qwen3-5-vlm-using-nvidia-gpu-accelerated-endpoints/" />
		<id>https://developer.nvidia.com/blog/?p=112969</id>
		<updated>2026-03-05T19:46:08Z</updated>
		<published>2026-02-27T17:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Mixture of Experts (MoE)" /><category scheme="https://developer.nvidia.com/blog" term="NIM" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" /><category scheme="https://developer.nvidia.com/blog" term="VLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="qwen3-5" />Alibaba has introduced the new open source Qwen3.5 series built for native multimodal agents. The first model in this series is a ~400B parameter native...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/develop-native-multimodal-agents-with-qwen3-5-vlm-using-nvidia-gpu-accelerated-endpoints/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="qwen3-5" />Alibaba has introduced the new open source Qwen3.5 series built for native multimodal agents. The first model in this series is a ~400B parameter native...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/qwen3-5-png.webp 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" title="qwen3-5" /><p>Alibaba has introduced the new open source Qwen3.5 series built for native multimodal agents. The first model in this series is a ~400B parameter native vision-language model (VLM) with reasoning built with a hybrid architecture of mixture of experts (MoE) and Gated Delta Networks. Qwen3.5 can understand and navigate user interfaces, which improves on the previous generation of VLMs. Qwen3.5…</p>
<p><a href="https://developer.nvidia.com/blog/develop-native-multimodal-agents-with-qwen3-5-vlm-using-nvidia-gpu-accelerated-endpoints/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
<link href="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/Qwen35.mp4" rel="enclosure" length="5261815" type="video/mp4" />
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/develop-native-multimodal-agents-with-qwen3-5-vlm-using-nvidia-gpu-accelerated-endpoints/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/develop-native-multimodal-agents-with-qwen3-5-vlm-using-nvidia-gpu-accelerated-endpoints/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Shwetha Krishnamurthy</name>
					</author>
		<title type="html"><![CDATA[Maximizing GPU Utilization with NVIDIA Run:ai and NVIDIA NIM]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/maximizing-gpu-utilization-with-nvidia-runai-and-nvidia-nim/" />
		<id>https://developer.nvidia.com/blog/?p=112973</id>
		<updated>2026-03-05T19:19:47Z</updated>
		<published>2026-02-27T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929.webp 1917w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-visual-mixture-of-experts-3105423" />Organizations deploying LLMs are challenged by inference workloads with different resource requirements. A small embedding model might use only a few gigabytes...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/maximizing-gpu-utilization-with-nvidia-runai-and-nvidia-nim/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929.webp 1917w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-visual-mixture-of-experts-3105423" />Organizations deploying LLMs are challenged by inference workloads with different resource requirements. A small embedding model might use only a few gigabytes...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-768x432.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-179x101.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-300x169.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-625x351.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1536x864.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-645x363.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-658x370.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-500x281.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-160x90.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-362x204.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-196x110.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-1024x576.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929-960x540.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/genai-visual-mixture-of-experts-3105423-e1772062206929.webp 1917w" sizes="auto, (max-width: 768px) 100vw, 768px" title="genai-visual-mixture-of-experts-3105423" /><p>Organizations deploying LLMs are challenged by inference workloads with different resource requirements. A small embedding model might use only a few gigabytes of GPU memory, while a 70B+ parameter LLM could require multiple GPUs. This diversity often leads to low average GPU utilization, high compute costs, and unpredictable latency. The problem isn’t just about packing more workloads onto…</p>
<p><a href="https://developer.nvidia.com/blog/maximizing-gpu-utilization-with-nvidia-runai-and-nvidia-nim/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/maximizing-gpu-utilization-with-nvidia-runai-and-nvidia-nim/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/maximizing-gpu-utilization-with-nvidia-runai-and-nvidia-nim/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Jamie Li</name>
					</author>
		<title type="html"><![CDATA[Making Softmax More Efficient with NVIDIA Blackwell Ultra]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/making-softmax-more-efficient-with-nvidia-blackwell-ultra/" />
		<id>https://developer.nvidia.com/blog/?p=112900</id>
		<updated>2026-03-05T19:19:48Z</updated>
		<published>2026-02-25T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell Ultra" /><category scheme="https://developer.nvidia.com/blog" term="cuDNN" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="GB200" /><category scheme="https://developer.nvidia.com/blog" term="GB300" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Tensor Cores" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" />LLM context lengths are exploding, and architectures are moving toward complex attention schemes like Multi-Head Latent Attention (MLA) and Grouped Query...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/making-softmax-more-efficient-with-nvidia-blackwell-ultra/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" />LLM context lengths are exploding, and architectures are moving toward complex attention schemes like Multi-Head Latent Attention (MLA) and Grouped Query...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/image2-4-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="image2" /><p>LLM context lengths are exploding, and architectures are moving toward complex attention schemes like Multi-Head Latent Attention (MLA) and Grouped Query Attention (GQA). As a result, AI ”speed of thought” is increasingly governed not by the massive throughput of matrix multiplications, but by the transcendental math of the softmax function. Transcendentals refer to functions that cannot be…</p>
<p><a href="https://developer.nvidia.com/blog/making-softmax-more-efficient-with-nvidia-blackwell-ultra/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/making-softmax-more-efficient-with-nvidia-blackwell-ultra/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/making-softmax-more-efficient-with-nvidia-blackwell-ultra/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Aditya Vavre</name>
					</author>
		<title type="html"><![CDATA[Using NVFP4 Low-Precision Model Training for Higher Throughput Without Losing Accuracy]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/using-nvfp4-low-precision-model-training-for-higher-throughput-without-losing-accuracy/" />
		<id>https://developer.nvidia.com/blog/?p=112818</id>
		<updated>2026-03-05T19:19:49Z</updated>
		<published>2026-02-23T18:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="NVFP4" /><category scheme="https://developer.nvidia.com/blog" term="Training AI Models" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-jpg.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="best-models-trained" />As the sizes of AI models and datasets continue to increase, relying only on higher-precision BF16 training is no longer sufficient. Key challenges such as...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/using-nvfp4-low-precision-model-training-for-higher-throughput-without-losing-accuracy/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-jpg.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="best-models-trained" />As the sizes of AI models and datasets continue to increase, relying only on higher-precision BF16 training is no longer sufficient. Key challenges such as...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1536x864.jpg 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-196x110.jpg 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-1024x576.jpg 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/best-models-trained-jpg.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="best-models-trained" /><p>As the sizes of AI models and datasets continue to increase, relying only on higher-precision BF16 training is no longer sufficient. Key challenges such as training throughput expectations, memory limits, and rising costs are becoming the primary barriers to scaling transformer models. Using lower-precision training can address these challenges. By reducing the numeric precision used during…</p>
<p><a href="https://developer.nvidia.com/blog/using-nvfp4-low-precision-model-training-for-higher-throughput-without-losing-accuracy/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/using-nvfp4-low-precision-model-training-for-higher-throughput-without-losing-accuracy/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/using-nvfp4-low-precision-model-training-for-higher-throughput-without-losing-accuracy/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Mukul Joshi</name>
					</author>
		<title type="html"><![CDATA[Accelerating Data Processing with NVIDIA Multi-Instance GPU and Locality Domains]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/accelerating-data-processing-with-nvidia-multi-instance-gpu-and-numa-node-localization/" />
		<id>https://developer.nvidia.com/blog/?p=112599</id>
		<updated>2026-04-10T23:11:35Z</updated>
		<published>2026-02-19T17:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="CUDA C++" /><category scheme="https://developer.nvidia.com/blog" term="Data Analytics / Processing" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Memory" /><category scheme="https://developer.nvidia.com/blog" term="Multi-Instance GPU (MIG)" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-jpg.webp 979w" sizes="auto, (max-width: 768px) 100vw, 768px" title="multicolored-bulging-cube" />NVIDIA flagship data center GPUs in the NVIDIA Ampere, NVIDIA Hopper, and NVIDIA Blackwell families all feature non-uniform memory access (NUMA) behaviors, but...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/accelerating-data-processing-with-nvidia-multi-instance-gpu-and-numa-node-localization/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-jpg.webp 979w" sizes="auto, (max-width: 768px) 100vw, 768px" title="multicolored-bulging-cube" />NVIDIA flagship data center GPUs in the NVIDIA Ampere, NVIDIA Hopper, and NVIDIA Blackwell families all feature non-uniform memory access (NUMA) behaviors, but...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-768x432.jpg 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-179x101.jpg 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-300x169.jpg 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-625x352.jpg 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-645x363.jpg 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-660x370.jpg 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-500x281.jpg 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-160x90.jpg 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-362x204.jpg 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-195x110.jpg 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-960x540.jpg 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multicolored-bulging-cube-1-jpg.webp 979w" sizes="auto, (max-width: 768px) 100vw, 768px" title="multicolored-bulging-cube" /><p>NVIDIA flagship data center GPUs in the NVIDIA Ampere, NVIDIA Hopper, and NVIDIA Blackwell families all feature non-uniform memory access (NUMA) behaviors, but expose a single memory space. Most programs therefore do not have an issue with memory non-uniformity. However, as bandwidth increases in newer generation GPUs, there are significant performance and power gains to be had when taking into…</p>
<p><a href="https://developer.nvidia.com/blog/accelerating-data-processing-with-nvidia-multi-instance-gpu-and-numa-node-localization/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/accelerating-data-processing-with-nvidia-multi-instance-gpu-and-numa-node-localization/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/accelerating-data-processing-with-nvidia-multi-instance-gpu-and-numa-node-localization/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Boskey Savla</name>
					</author>
		<title type="html"><![CDATA[Unlock Massive Token Throughput with GPU Fractioning in NVIDIA Run:ai]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/unlock-massive-token-throughput-with-gpu-fractioning-in-nvidia-runai/" />
		<id>https://developer.nvidia.com/blog/?p=112734</id>
		<updated>2026-03-05T19:19:51Z</updated>
		<published>2026-02-18T18:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-300x169-png.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-625x352-png.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-179x101-png.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1536x864-png.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-645x363-png.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-660x370-png.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-500x281-png.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-160x90-png.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-362x204-png.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-196x110-png.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1024x576-png.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-960x540-png.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="run ai featured" />As AI workloads scale, achieving high throughput, efficient resource usage, and predictable latency becomes essential. NVIDIA Run:ai addresses these challenges...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/unlock-massive-token-throughput-with-gpu-fractioning-in-nvidia-runai/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-300x169-png.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-625x352-png.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-179x101-png.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1536x864-png.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-645x363-png.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-660x370-png.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-500x281-png.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-160x90-png.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-362x204-png.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-196x110-png.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1024x576-png.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-960x540-png.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="run ai featured" />As AI workloads scale, achieving high throughput, efficient resource usage, and predictable latency becomes essential. NVIDIA Run:ai addresses these challenges...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-768x432-png.webp 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-300x169-png.webp 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-625x352-png.webp 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-179x101-png.webp 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1536x864-png.webp 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-645x363-png.webp 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-660x370-png.webp 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-500x281-png.webp 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-160x90-png.webp 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-362x204-png.webp 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-196x110-png.webp 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-1024x576-png.webp 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-960x540-png.webp 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2025/06/run-ai-featured-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="run ai featured" /><p>As AI workloads scale, achieving high throughput, efficient resource usage, and predictable latency becomes essential. NVIDIA Run:ai addresses these challenges through intelligent scheduling and dynamic GPU fractioning. GPU fractioning is wholly delivered by NVIDIA Run:ai in any environment—cloud, NCP, and on-premises. This post presents the joint benchmarking effort between NVIDIA and AI…</p>
<p><a href="https://developer.nvidia.com/blog/unlock-massive-token-throughput-with-gpu-fractioning-in-nvidia-runai/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/unlock-massive-token-throughput-with-gpu-fractioning-in-nvidia-runai/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/unlock-massive-token-throughput-with-gpu-fractioning-in-nvidia-runai/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Daniel Rodriguez</name>
					</author>
		<title type="html"><![CDATA[Topping the GPU MODE Kernel Leaderboard with NVIDIA cuda.compute]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/topping-the-gpu-mode-kernel-leaderboard-with-nvidia-cuda-compute/" />
		<id>https://developer.nvidia.com/blog/?p=112718</id>
		<updated>2026-03-05T19:19:52Z</updated>
		<published>2026-02-18T17:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="CUDA" /><category scheme="https://developer.nvidia.com/blog" term="featured" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cuda compute gpu mode" />Python dominates machine learning for its ergonomics, but writing truly fast GPU code has historically meant dropping into C++ to write custom kernels and to...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/topping-the-gpu-mode-kernel-leaderboard-with-nvidia-cuda-compute/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cuda compute gpu mode" />Python dominates machine learning for its ergonomics, but writing truly fast GPU code has historically meant dropping into C++ to write custom kernels and to...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/cuda-compute-gpu-mode-png.webp 1600w" sizes="auto, (max-width: 768px) 100vw, 768px" title="cuda compute gpu mode" /><p>Python dominates machine learning for its ergonomics, but writing truly fast GPU code has historically meant dropping into C++ to write custom kernels and to maintain bindings back to Python. For most Python developers and researchers, this is a significant barrier to entry. Frameworks like PyTorch address this by implementing kernels in CUDA C++—either handwritten or by leveraging libraries…</p>
<p><a href="https://developer.nvidia.com/blog/topping-the-gpu-mode-kernel-leaderboard-with-nvidia-cuda-compute/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/topping-the-gpu-mode-kernel-leaderboard-with-nvidia-cuda-compute/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/topping-the-gpu-mode-kernel-leaderboard-with-nvidia-cuda-compute/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Utkarsh Uppal</name>
					</author>
		<title type="html"><![CDATA[How NVIDIA Extreme Hardware-Software Co-Design Delivered a Large Inference Boost for Sarvam AI’s Sovereign Models]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/" />
		<id>https://developer.nvidia.com/blog/?p=112699</id>
		<updated>2026-03-05T19:19:53Z</updated>
		<published>2026-02-18T16:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="Data Science" /><category scheme="https://developer.nvidia.com/blog" term="Blackwell" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="H100" /><category scheme="https://developer.nvidia.com/blog" term="NeMo" /><category scheme="https://developer.nvidia.com/blog" term="Nemotron" /><category scheme="https://developer.nvidia.com/blog" term="NVIDIA Inception" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-dgx-cloud-ari-blog-1920x1080" />As global AI adoption accelerates, developers face a growing challenge: delivering large language model (LLM) performance that meets real-world latency and cost...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-dgx-cloud-ari-blog-1920x1080" />As global AI adoption accelerates, developers face a growing challenge: delivering large language model (LLM) performance that meets real-world latency and cost...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/ov-dgx-cloud-ari-blog-1920x1080-2-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="ov-dgx-cloud-ari-blog-1920x1080" /><p>As global AI adoption accelerates, developers face a growing challenge: delivering large language model (LLM) performance that meets real-world latency and cost requirements. Running models with tens of billions of parameters in production, especially for conversational or voice-based AI agents, demands high throughput, low latency, and predictable service-level performance.</p>
<p><a href="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Shruthii Sathyanarayanan</name>
					</author>
		<title type="html"><![CDATA[Build AI-Ready Knowledge Systems Using 5 Essential Multimodal RAG Capabilities]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/build-ai-ready-knowledge-systems-using-5-essential-multimodal-rag-capabilities/" />
		<id>https://developer.nvidia.com/blog/?p=112325</id>
		<updated>2026-03-13T21:02:27Z</updated>
		<published>2026-02-17T18:00:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Data Center / Cloud" /><category scheme="https://developer.nvidia.com/blog" term="AI Agent" /><category scheme="https://developer.nvidia.com/blog" term="AI Data Platform" /><category scheme="https://developer.nvidia.com/blog" term="AI-Ready Data" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="Retrieval Augmented Generation (RAG)" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="retrieval-augmented-generation" />Enterprise data is inherently complex: real-world documents are multimodal, spanning text, tables, charts and graphs, images, diagrams, scanned pages, forms,...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/build-ai-ready-knowledge-systems-using-5-essential-multimodal-rag-capabilities/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="retrieval-augmented-generation" />Enterprise data is inherently complex: real-world documents are multimodal, spanning text, tables, charts and graphs, images, diagrams, scanned pages, forms,...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-196x110.png 196w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/retrieval-augmented-generation-png.webp 1920w" sizes="auto, (max-width: 768px) 100vw, 768px" title="retrieval-augmented-generation" /><p>Enterprise data is inherently complex: real-world documents are multimodal, spanning text, tables, charts and graphs, images, diagrams, scanned pages, forms, and embedded metadata. Financial reports carry critical insights in tables, engineering manuals rely on diagrams, and legal documents often include annotated or scanned content. Retrieval-augmented generation (RAG) was created to ground…</p>
<p><a href="https://developer.nvidia.com/blog/build-ai-ready-knowledge-systems-using-5-essential-multimodal-rag-capabilities/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/build-ai-ready-knowledge-systems-using-5-essential-multimodal-rag-capabilities/#comments" thr:count="1"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/build-ai-ready-knowledge-systems-using-5-essential-multimodal-rag-capabilities/feed/" thr:count="1"/>
		<thr:total>1</thr:total>
	</entry>
		<entry>
		<author>
			<name>Oyindamola Omotuyi</name>
					</author>
		<title type="html"><![CDATA[R²D²: Scaling Multimodal Robot Learning with NVIDIA Isaac Lab]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/r2d2-scaling-multimodal-robot-learning-with-nvidia-isaac-lab/" />
		<id>https://developer.nvidia.com/blog/?p=112456</id>
		<updated>2026-03-05T19:19:55Z</updated>
		<published>2026-02-10T18:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Robotics" /><category scheme="https://developer.nvidia.com/blog" term="Simulation / Modeling / Design" /><category scheme="https://developer.nvidia.com/blog" term="Top Stories" /><category scheme="https://developer.nvidia.com/blog" term="AI Foundation Models" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Humanoid Robots" /><category scheme="https://developer.nvidia.com/blog" term="NVIDIA Research" /><category scheme="https://developer.nvidia.com/blog" term="Open Source" /><category scheme="https://developer.nvidia.com/blog" term="Physical AI" /><category scheme="https://developer.nvidia.com/blog" term="Robotics Research and Development Digest (R²D²)" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="multimodal-robotics" />Building robust, intelligent robots requires testing them in complex environments. However, gathering data in the physical world is expensive, slow, and often...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/r2d2-scaling-multimodal-robot-learning-with-nvidia-isaac-lab/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="multimodal-robotics" />Building robust, intelligent robots requires testing them in complex environments. However, gathering data in the physical world is expensive, slow, and often...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1.gif 600w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-179x101.gif 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-300x169.gif 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-500x282.gif 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-160x90.gif 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-362x204.gif 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/multimodal-robotics-1-195x110.gif 195w" sizes="auto, (max-width: 600px) 100vw, 600px" title="multimodal-robotics" /><p>Building robust, intelligent robots requires testing them in complex environments. However, gathering data in the physical world is expensive, slow, and often dangerous. It is nearly impossible to safely train for real-world critical risks, such as high-speed collisions or hardware failures. Worse, real-world data is usually biased toward “normal” conditions, leaving robots unprepared for the…</p>
<p><a href="https://developer.nvidia.com/blog/r2d2-scaling-multimodal-robot-learning-with-nvidia-isaac-lab/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/r2d2-scaling-multimodal-robot-learning-with-nvidia-isaac-lab/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/r2d2-scaling-multimodal-robot-learning-with-nvidia-isaac-lab/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>Quynh L. Nguyen</name>
					</author>
		<title type="html"><![CDATA[Using Accelerated Computing to Live-Steer Scientific Experiments at Massive Research Facilities]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/using-accelerated-computing-to-live-steer-scientific-experiments-at-massive-research-facilities/" />
		<id>https://developer.nvidia.com/blog/?p=110460</id>
		<updated>2026-03-05T19:19:56Z</updated>
		<published>2026-02-10T17:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Edge Computing" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="research" />		<summary type="html"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/star-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="star-gif" />Scientists and engineers who design and build unique scientific research facilities face similar challenges. These include managing massive data rates that...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/using-accelerated-computing-to-live-steer-scientific-experiments-at-massive-research-facilities/"><![CDATA[<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/star-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="star-gif" />Scientists and engineers who design and build unique scientific research facilities face similar challenges. These include managing massive data rates that...<img width="600" height="338" src="https://developer-blogs.nvidia.com/wp-content/uploads/2025/12/star-gif.gif" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" title="star-gif" /><p>Scientists and engineers who design and build unique scientific research facilities face similar challenges. These include managing massive data rates that exceed current computational infrastructure capacity to extract scientific insights and driving the experiments in real time. These challenges are obstacles to maximizing the impact of scientific discoveries and significantly slow the pace of…</p>
<p><a href="https://developer.nvidia.com/blog/using-accelerated-computing-to-live-steer-scientific-experiments-at-massive-research-facilities/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/using-accelerated-computing-to-live-steer-scientific-experiments-at-massive-research-facilities/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/using-accelerated-computing-to-live-steer-scientific-experiments-at-massive-research-facilities/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
		<entry>
		<author>
			<name>​​Lucas Liebenwein</name>
					</author>
		<title type="html"><![CDATA[Automating Inference Optimizations with NVIDIA TensorRT LLM AutoDeploy]]></title>
		<link rel="alternate" type="text/html" href="https://developer.nvidia.com/blog/automating-inference-optimizations-with-nvidia-tensorrt-llm-autodeploy/" />
		<id>https://developer.nvidia.com/blog/?p=112441</id>
		<updated>2026-03-05T19:19:57Z</updated>
		<published>2026-02-09T18:30:00Z</published>
		<category scheme="https://developer.nvidia.com/blog" term="Agentic AI / Generative AI" /><category scheme="https://developer.nvidia.com/blog" term="Developer Tools &amp; Techniques" /><category scheme="https://developer.nvidia.com/blog" term="MLOps" /><category scheme="https://developer.nvidia.com/blog" term="AI Inference" /><category scheme="https://developer.nvidia.com/blog" term="featured" /><category scheme="https://developer.nvidia.com/blog" term="Inference Performance" /><category scheme="https://developer.nvidia.com/blog" term="LLMs" /><category scheme="https://developer.nvidia.com/blog" term="PyTorch" />		<summary type="html"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-png.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="llm-optimize-deploy" />NVIDIA TensorRT LLM enables developers to build high-performance inference engines for large language models (LLMs), but deploying a new architecture...]]></summary>
		<content type="html" xml:base="https://developer.nvidia.com/blog/automating-inference-optimizations-with-nvidia-tensorrt-llm-autodeploy/"><![CDATA[<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-png.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="llm-optimize-deploy" />NVIDIA TensorRT LLM enables developers to build high-performance inference engines for large language models (LLMs), but deploying a new architecture...<img width="768" height="432" src="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png" class="webfeedsFeaturedVisual wp-post-image" alt="" style="display: block; margin-bottom: 5px; clear:both;max-width: 100%;" link_thumbnail="" decoding="async" loading="lazy" srcset="https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-768x432.png 768w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-179x101.png 179w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-300x169.png 300w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-625x352.png 625w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1536x864.png 1536w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-645x363.png 645w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-660x370.png 660w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-500x281.png 500w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-160x90.png 160w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-362x204.png 362w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-195x110.png 195w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-1024x576.png 1024w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-960x540.png 960w, https://developer-blogs.nvidia.com/wp-content/uploads/2026/02/llm-optimize-deploy-1-png.webp 1999w" sizes="auto, (max-width: 768px) 100vw, 768px" title="llm-optimize-deploy" /><p>NVIDIA TensorRT LLM enables developers to build high-performance inference engines for large language models (LLMs), but deploying a new architecture traditionally requires significant manual effort. To address this challenge, today we are announcing the availability of AutoDeploy as a beta feature in TensorRT LLM. AutoDeploy compiles off-the-shelf PyTorch models into inference-optimized…</p>
<p><a href="https://developer.nvidia.com/blog/automating-inference-optimizations-with-nvidia-tensorrt-llm-autodeploy/" rel="nofollow" data-wpel-link="internal" target="_self">Source</a></p>]]></content>
		<link rel="replies" type="text/html" href="https://developer.nvidia.com/blog/automating-inference-optimizations-with-nvidia-tensorrt-llm-autodeploy/#comments" thr:count="0"/>
		<link rel="replies" type="application/atom+xml" href="https://developer.nvidia.com/blog/automating-inference-optimizations-with-nvidia-tensorrt-llm-autodeploy/feed/" thr:count="0"/>
		<thr:total>0</thr:total>
	</entry>
	</feed>