{"id":91552,"date":"2025-08-18T09:52:12","date_gmt":"2025-08-18T09:52:12","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/91552\/"},"modified":"2025-08-18T09:52:12","modified_gmt":"2025-08-18T09:52:12","slug":"nvidia-opens-portals-to-world-of-robotics-with-new-omniverse-libraries-cosmos-physical-ai-models-and-ai-computing-infrastructure-2","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/91552\/","title":{"rendered":"NVIDIA Opens Portals to World of Robotics With New Omniverse Libraries, Cosmos Physical AI Models and AI Computing Infrastructure"},"content":{"rendered":"<p>         &#13;<br \/>\n\tNew NVIDIA\u00a0Omniverse\u00a0NuRec\u00a03D Gaussian Splatting Libraries Enable Large-Scale World Reconstruction&#13;<br \/>\n\tNew NVIDIA Cosmos\u00a0Models\u00a0Enable World Generation and Spatial Reasoning&#13;<br \/>\n\tNew NVIDIA\u00a0RTX\u00a0PRO Blackwell Servers and NVIDIA DGX Cloud Let Developers Run the Most Demanding Simulations Anywhere&#13;<br \/>\n\tPhysical AI Leaders Amazon Devices &amp; Services, Boston Dynamics, Figure AI and Hexagon Embrace Simulation and Synthetic Data Generation&#13;<\/p>\n<p>SIGGRAPH\u2014NVIDIA today announced new <a href=\"https:\/\/www.nvidia.com\/en-us\/omniverse\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA Omniverse<\/a>\u2122 libraries and <a href=\"https:\/\/www.nvidia.com\/en-us\/ai\/cosmos\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA Cosmos<\/a>\u2122 world foundation models (<a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/world-models\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">WFMs<\/a>) that accelerate the development and deployment of robotics solutions.<\/p>\n<p>Powered by new <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/rtx-pro-6000-blackwell-server-edition\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA RTX PRO\u2122 Servers<\/a> and <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-cloud\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA DGX\u2122 Cloud<\/a>, the libraries and models let developers anywhere develop physically accurate <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/digital-twin\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">digital twins<\/a>, capture and reconstruct the real world in simulation, <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-synthetic-data\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">generate synthetic data<\/a> for training <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/generative-physical-ai\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">physical AI<\/a> models and build AI agents that understand the physical world.<\/p>\n<p>\u201cComputer graphics and AI are converging to fundamentally transform robotics,\u201d said Rev Lebaredian, vice president of Omniverse and simulation technologies at NVIDIA. \u201cBy combining AI reasoning with scalable, physically accurate simulation, we\u2019re enabling developers to build tomorrow\u2019s robots and autonomous vehicles that will transform trillions of dollars in industries.\u201d<\/p>\n<p>New NVIDIA Omniverse Libraries Advance Applications for World Composition<br \/>&#13;<br \/>\nNew <a href=\"https:\/\/developer.nvidia.com\/omniverse\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA Omniverse software development kits (SDKs) and libraries<\/a> are <a href=\"https:\/\/developer.nvidia.com\/blog\/developers-build-fast-and-reliable-robot-simulations-with-nvidia-omniverse-libraries\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">now available<\/a> for building and deploying <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/industrial-ai\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">industrial AI<\/a> and <a href=\"https:\/\/www.nvidia.com\/en-us\/use-cases\/robotics-simulation\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">robotics simulation<\/a> applications.<\/p>\n<p><a href=\"https:\/\/carla.org\/2025\/06\/11\/release-0.9.16-pre\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Omniverse NuRec rendering is now integrated in CARLA<\/a>, a leading open-source simulator used by over 150,000 developers. <a href=\"https:\/\/www.nvidia.com\/en-us\/solutions\/autonomous-vehicles\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Autonomous vehicle<\/a> (AV) toolchain leader\u00a0<a href=\"https:\/\/www.foretellix.com\/scalable-neural-reconstruction-av\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Foretellix<\/a> is integrating NuRec, <a href=\"https:\/\/www.foretellix.com\/data-automation-toolchain-for-ai-powered-av-development\/\" rel=\"nofollow noopener\" target=\"\" title=\"\">NVIDIA Omniverse Sensor RTX<\/a>\u2122 and Cosmos Transfer to enhance its scalable synthetic data generation with physically accurate scenarios. <a href=\"https:\/\/voxel51.com\/blog\/enabling-av-datasets-nvidia-nurec-and-fiftyone?utm_source=Referral&amp;utm_medium=Nvidia&amp;utm_campaign=Nvidia_Nurec_announcement_SIGGRAPH_conference\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Voxel51\u2019s<\/a> data engine for visual and multimodal AI, FiftyOne, supports NuRec to ease data preparation for reconstructions. FiftyOne is used by customers such as Ford and Porsche.<\/p>\n<p>Boston Dynamics, Figure AI, <a href=\"https:\/\/s2025.conference-schedule.org\/presentation\/?id=ind_114&amp;sess=sess481\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Hexagon<\/a>, RAI Institute, <a href=\"https:\/\/www.nvidia.com\/en-us\/customer-stories\/lightwheel\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Lightwheel<\/a> and Skild AI are adopting Omniverse libraries, Isaac Sim and <a href=\"https:\/\/developer.nvidia.com\/isaac\/lab\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Isaac Lab<\/a> to accelerate their AI robotics development, while <a href=\"https:\/\/blogs.nvidia.com\/blog\/amazon-zero-touch-manufacturing\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Amazon Devices &amp; Services<\/a> is using them to power a new manufacturing solution.<\/p>\n<p>Cosmos Advances World Generation for Robotics<br \/>&#13;<br \/>\nCosmos WFMs, downloaded over 2 million times, let developers generate diverse data for training robots at scale using text, image and video prompts.<\/p>\n<p><a href=\"https:\/\/www.google.com\/url?q=https:\/\/github.com\/nvidia-cosmos&amp;sa=D&amp;source=docs&amp;ust=1753918585556593&amp;usg=AOvVaw0p0Hvd1WuBtn38793ozugb\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">New models<\/a> announced at SIGGRAPH deliver major advances in synthetic data generation speed, accuracy, language support and control:<\/p>\n<p>&#13;<br \/>\n\t<a href=\"https:\/\/github.com\/nvidia-cosmos\/cosmos-transfer2\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">Cosmos Transfer-2<\/a>, coming soon, simplifies prompting and accelerates photorealistic synthetic data generation from ground-truth 3D simulation scenes or spatial control inputs like depth, segmentation, edges and high-definition maps.&#13;<br \/>\n\tA distilled version of Cosmos Transfer reduces the 70-step distillation process to one so developers can run the model on <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/products\/rtx-pro-server\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA RTX PRO Servers<\/a> at unprecedented speed.&#13;<\/p>\n<p>Lightwheel, Moon Surgical and Skild AI are using Cosmos Transfer to accelerate physical AI training by simulating diverse conditions at scale.<\/p>\n<p>Cosmos Reason Breaks Through World Understanding<br \/>&#13;<br \/>\nSince the introduction of OpenAI\u2019s CLIP model, <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/vision-language-models\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">vision language models<\/a> (VLMs) have transformed computer-vision tasks like object and pattern recognition. However, they have not yet been able to solve multistep tasks nor handle ambiguity or novel experiences.<\/p>\n<p><a href=\"https:\/\/build.nvidia.com\/nvidia\/cosmos-reason1-7b\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA Cosmos Reason<\/a> \u2014 a new open, customizable, 7-billion-parameter reasoning VLM for physical AI and robotics \u2014 lets robots and vision AI agents reason like humans, using prior knowledge, physics understanding and common sense to understand and act in the real world.\u00a0<\/p>\n<p>Cosmos Reason can be used for robotics and physical AI applications including:<\/p>\n<p>NVIDIA\u2019s robotics and <a href=\"https:\/\/developer.nvidia.com\/drive\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA DRIVE<\/a>\u2122 teams are using Cosmos Reason for data curation and filtering, annotation and VLA post-training. Uber is using it to annotate and caption AV training data.<\/p>\n<p><a href=\"https:\/\/www.magna.com\/stories\/blog\/2025\/magna-accelerates-operational-ai-using-a-broad-ecosystem--including-nvidia-ominverse-and-cosmos\" rel=\"nofollow noopener\" target=\"_blank\" title=\"Magna\">Magna<\/a>\u00a0is developing with Cosmos Reason as part of its City Delivery platform \u2014 a fully autonomous, low-cost solution for instant delivery \u2014 to help vehicles adapt more quickly to new cities. Cosmos Reason adds world understanding to the vehicles\u2019 long-term trajectory planner. <a href=\"https:\/\/www.vastdata.com\/blog\/unified-ai-for-urban-intelligence-public-safety\" rel=\"nofollow noopener\" target=\"_blank\" title=\"VAST Data\">VAST Data<\/a>, Milestone Systems and\u00a0<a href=\"https:\/\/www.linkervision.com\/post\/enable-smart-cities-with-physical-ai%E2%80%94driven-by-nvidia-cosmos-reason\" rel=\"nofollow noopener\" target=\"_blank\" title=\"Linker Vision\">Linker Vision<\/a> are adopting Cosmos Reason to automate traffic monitoring, improve safety and enhance visual inspection in cities and industrial settings.<\/p>\n<p>New NVIDIA AI Infrastructure Powers Robotics Workloads Anywhere<br \/>&#13;<br \/>\nTo enable developers to take full advantage of these advanced technologies and software libraries, NVIDIA announced AI infrastructure designed for the most demanding workloads.<\/p>\n<p>Accelerating the Developer Ecosystem<br \/>&#13;<br \/>\nTo help robotics and physical AI developers advance 3D and simulation technology adoption, NVIDIA also announced:<\/p>\n<p>&#13;<br \/>\n\tOpenUSD Curriculum and Certification, which addresses demand for USD expertise, with support from\u00a0<a href=\"https:\/\/aousd.org\/uncategorized\/alliance-for-openusd-announces-new-members-inclusive-language-guide-and-core-specification-progress\/\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">AOUSD<\/a> members Adobe, Amazon Robotics, Ansys \u2014 part of Synopsys, Autodesk, Pixar, PTC, Rockwell Automation, SideFX, Siemens, TCS and Trimble, as well as industry leaders such as Hexagon.&#13;<br \/>\n\tOpen-source collaboration with Lightwheel to integrate robot policy training and evaluation frameworks into NVIDIA Isaac Lab, featuring parallel reinforcement learning training capabilities, benchmarks and simulation-ready assets for robot manipulation and locomotion.&#13;<\/p>\n<p>Watch the <a href=\"https:\/\/www.youtube.com\/watch?v=rFcmv2pXR0w\" rel=\"nofollow noopener\" target=\"_blank\" title=\"\">NVIDIA Research special address at SIGGRAPH<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":"&#13; New NVIDIA\u00a0Omniverse\u00a0NuRec\u00a03D Gaussian Splatting Libraries Enable Large-Scale World Reconstruction&#13; New NVIDIA Cosmos\u00a0Models\u00a0Enable World Generation and Spatial Reasoning&#13;&hellip;\n","protected":false},"author":2,"featured_media":91553,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[46],"tags":[191,74],"class_list":{"0":"post-91552","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-computing","9":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/91552","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=91552"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/91552\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/91553"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=91552"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=91552"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=91552"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}