Category: Artificial intelligence

  • OpenAI Acquires Jony Ive’s AI Gadget for $6.5 Billion

    OpenAI Acquires Jony Ive’s AI Gadget for $6.5 Billion

    Key Takeaways

    1. OpenAI has partnered with Jony Ive’s design studio, LoveFrom, to create a new entity called io for AI hardware development.
    2. OpenAI is investing $6.5 billion in shares to acquire io, despite the startup not yet having products or revenue.
    3. The collaboration includes a team of about 55 engineers and developers, and a prototype is already being tested by Sam Altman.
    4. io’s first product, aimed to complement smartphones, is expected to launch in 2026, amidst challenges from past AI device failures.
    5. LoveFrom will help OpenAI improve the user interface of ChatGPT as part of the partnership.


    OpenAI has recently revealed its partnership with Jony Ive’s design studio, LoveFrom, to establish a new entity named io, which aims to produce AI hardware. According to a report by Bloomberg, this collaboration is more than just a simple partnership, as OpenAI is allegedly investing $6.5 billion in shares to acquire io.

    A Bold Move

    The startup has not yet introduced any products to the market or earned any revenue, making this investment seem quite astonishing. This acquisition will provide OpenAI with a focused hardware team comprised of about 55 engineers and software developers. Additionally, it will foster collaboration with Jony Ive and his team, along with a hardware product that is likely already in advanced stages of development. Jony Ive mentioned in an embedded video that Sam Altman, the head of OpenAI, has even taken a prototype home for testing.

    Future Challenges

    Despite the promising partnership, after the notable failures of AI devices like the Humane Ai Pin in recent years, it remains uncertain if OpenAI and LoveFrom can effectively compete against smartphones and AI applications. The upcoming device is designed to complement smartphones rather than replace them, aiming to provide users with a novel way to engage with AI. io’s first product is set to debut in 2026. Additionally, LoveFrom will assist OpenAI in revamping the user interface of ChatGPT.

    Source:
    Link

  • Prototype 2T1R Control Chip for Neuromorphic Computing Developed

    Prototype 2T1R Control Chip for Neuromorphic Computing Developed

    Key Takeaways

    1. In-memory computing is changing computer architecture by moving processing closer to memory for improved efficiency.
    2. A new 2T1R memristor design enhances energy efficiency for AI and edge devices by minimizing sneak path currents and leakage.
    3. The design supports analogue vector-matrix multiplication (VMM), crucial for machine learning applications.
    4. It effectively addresses virtual ground issues and wire resistance to boost performance and reduce power usage.
    5. This technology paves the way for faster, more integrated AI operations in memory, hinting at a future where hard drives could process information more intelligently.


    What if your hard drive did more than just store your files? Imagine it could think and react to information right where it’s located. This idea is part of in-memory computing, which is changing the way computers are built by moving processing closer to memory to improve efficiency.

    New Memristor Design

    Researchers from Forschungszentrum Jülich and the University of Duisburg-Essen have introduced a fresh design based on a 2T1R memristor. This innovation aims to facilitate a more energy-efficient approach for AI and edge devices.

    The details of this design, shared on arXiv, include the integration of two transistors and one memristor in each cell, which helps regulate current to minimize sneak path currents—an issue often faced in memristor arrays. Unike regular memory, this new design connects both memristor terminals to ground when not in active use. This method could enhance signal stability and cut down on leakage.

    Enhancing Machine Learning

    The architecture is meant to enable analogue vector-matrix multiplication (VMM), a key element in machine learning. It achieves this by managing memristor conductance with built-in DACs, PWM signals, and controlled current paths. A functional 2×2 test array was successfully created using standard 28 nm CMOS technology.

    By tackling virtual ground concerns and the impact of wire resistance, the design seeks to boost performance reliability and decrease power usage. Compatible with RISC-V control and digital connections, the 2T1R structure may set the stage for scalable neuromorphic chips, allowing for speedier and more compact AI acceleration directly in memory.

    Future of AI and Memory

    While your hard drive isn’t quite thinking yet, the technology that could make this a reality is already being developed in silicon. This progression hints at a future where AI operates faster and more seamlessly integrated with memory.

    Source:
    Link

  • Google Gemini 2.5 Update: Agent Mode, Deep Think, and Tools

    Google Gemini 2.5 Update: Agent Mode, Deep Think, and Tools

    Key Takeaways

    1. Gemini 2.5 Flash: Delivers improved reasoning and efficiency, operating 20-30% lighter in token consumption, available for all users and developers starting May 20.

    2. Deep Think Mode: Focuses on enhanced reasoning capabilities for math, coding, and multimodal evaluations, currently being tested by select API users with further safety evaluations planned.

    3. Agent Mode for Subscribers: Allows Gemini Ultra subscribers to set goals, pulling information from live searches and Google applications through a split interface.

    4. Gemini Canvas Enhancements: Introduces a “Create” menu for generating webpages, infographics, quizzes, audio summaries, and personalized app frameworks.

    5. Deep Research Tool: Enables users to merge public information with private documents, enhancing research capabilities for students and professionals.


    At I/O 2025, Google has unveiled a significant range of updates for its Gemini application. The firm revealed new features such as Gemini 2.5 Flash, an enhanced Deep Think mode, a ChatGPT-like Agent Mode, as well as support for quiz creation, audio summaries, and research based on documents—all integrated into a single platform.

    Key Upgrade: Gemini 2.5 Flash

    The most notable enhancement at this moment is Gemini 2.5 Flash, which offers improved reasoning abilities and greater efficiency. It operates 20-30% lighter in terms of token consumption and is already available in the Gemini app for all users. Developers and enterprise clients can also utilize an updated version via Google AI Studio and Vertex AI starting May 20. The complete Gemini 2.5 Pro model is anticipated to be released in early June.

    Deep Think Mode and Its Features

    In addition, Gemini 2.5 Deep Think is being promoted as an upgrade centered on reasoning, demonstrating better results in areas like math (USAMO 2025), coding (LiveCodeBench v6), and multimodal evaluations (MMMU). Google mentions this mode will undergo additional safety evaluations before being made public, but it is already being tested by a select group of Gemini API users.

    New Agent Mode for Subscribers

    If you are a Gemini Ultra subscriber, you will soon gain access to Agent Mode, which is powered by something known as Project Mariner. In this mode, you simply tell Gemini your goal, and it pulls information from live web searches, your Google applications, and external resources to accomplish it. This function features a split display where the chat interface is on the left, while a browser-like panel manages content on the right.

    On the creative front, Gemini Canvas has introduced a new “Create” menu, allowing users to generate complete webpages, infographics, quizzes, or audio summaries directly through chat. You can also describe a personalized app and let Gemini create a starting framework. Additionally, for students or working professionals, a Deep Research tool now enables you to merge public information with private PDFs, Google Drive documents, and market insights.

    Source:
    Link


  • Nvidia CEO Refutes Claims of AI Chip Diversion

    Nvidia CEO Refutes Claims of AI Chip Diversion

    Key Takeaways

    1. Nvidia’s CEO Jensen Huang stated there’s “no evidence of any AI chip diversion” to restricted areas, emphasizing the heavy and integrated design of their new data-center systems.
    2. Customers and governments are closely monitoring compliance with regulations to continue purchasing Nvidia products, and Huang supports the elimination of previous “AI diffusion” restrictions.
    3. There is rising demand for Nvidia accelerators in the Middle East, particularly from the UAE and Saudi Arabia, which Huang believes can be met without changing current production allocations.
    4. Critics highlight concerns over gray markets, with reports of Chinese buyers acquiring GPUs through shell corporations and illegal sales, prompting investigations in Singapore and calls for action from U.S. officials.
    5. Lawmakers in Washington are considering geo-tracking for high-end processors due to concerns about smuggling, but Nvidia argues that tracking individual components after shipment is nearly impossible.


    Nvidia’s CEO Jensen Huang made a stop in Taipei during Computex 2025, where he strongly stated that there’s “no evidence of any AI chip diversion” to restricted areas. In an interview with Bloomberg, he explained that data-center systems designed with the newly introduced Grace Blackwell architecture are quite heavy, nearly two tons, and are shipped as integrated racks, which makes it very difficult for any secret exports to happen.

    Monitoring and Compliance

    Huang also emphasized that both customers and governments are well aware of the regulations and “monitor themselves very carefully,” as they wish to continue purchasing Nvidia products. He mentioned that Washington’s choice to eliminate prior “AI diffusion” restrictions was a wise decision, aligning with his ongoing belief that limiting American technology abroad is fundamentally incorrect.

    Rising Demand in the Middle East

    These comments come at a time when the Middle East, especially the United Arab Emirates and Saudi Arabia, are looking for more Nvidia accelerators to support their local AI initiatives. Huang noted that effective production planning could meet this demand without needing to alter current allocations.

    Concerns Over Grey Markets

    However, critics argue that the situation is more complicated than it appears. Reports from this year indicate that Chinese buyers are acquiring H-class GPUs through shell corporations in Malaysia, Vietnam, and Taiwan; one reseller even showcased illegal H200 boards on social media. In response, Singapore has started investigations, and U.S. officials have urged Malaysian authorities to address what they describe as a rapidly growing gray market, which has seen imports of advanced GPUs surge more than 3,400 percent.

    Legislative Proposals

    In Washington, lawmakers are contemplating a requirement for manufacturers to implement geo-tracking on high-end gaming and AI processors, claiming that weight isn’t a sufficient deterrent to smuggling—pointing out that stolen cars frequently cross borders. Nvidia has responded by stating that once their servers leave the company, tracking individual components is nearly impossible, highlighting the disconnect between regulatory goals and actual technical capabilities.

    Source:
    Link

  • Nvidia and Foxconn to Build 10,000-GPU AI Supercomputer in Taiwan

    Nvidia and Foxconn to Build 10,000-GPU AI Supercomputer in Taiwan

    Key Takeaways

    1. Nvidia and Foxconn are partnering to create an “AI factory” supercomputer in southern Taiwan, integrating 10,000 Nvidia Blackwell GPUs.
    2. The project involves an investment of several hundred million US dollars, making it one of the largest AI investments globally, although smaller than Elon Musk’s Memphis Supercluster.
    3. The computing power will be distributed to local researchers, startups, and businesses to boost AI adoption in Taiwan.
    4. The initiative aims to create a larger AI ecosystem, linking public agencies and private enterprises, with goals to accelerate semiconductor innovation and enhance smart-city services.
    5. This supercomputer aligns with Nvidia’s expansion strategy, including plans for a research center in Shanghai and a fleet of AI servers in the U.S. worth half a trillion dollars.


    Nvidia and Foxconn have broadened their long-term partnership with new plans for an “AI factory” supercomputer in southern Taiwan. This system is set to be delivered by Big Innovation Company, a Foxconn subsidiary and Nvidia Cloud Partner, and will integrate 10,000 Nvidia Blackwell GPUs into one setup.

    Investment Overview

    With an investment of several hundred million US dollars, this project is less massive than Elon Musk’s Memphis Supercluster, which has 200,000 GPUs. However, it still stands as one of the largest AI investments globally. Nvidia is responsible for the hardware supply, whereas Foxconn will take care of the datacenter infrastructure and operational support.

    Local Impact

    The National Science and Technology Council of Taiwan plans to distribute the computing power of this cluster to local researchers, startups, and established businesses, thus boosting AI adoption domestically. Taiwan Semiconductor Manufacturing Company (TSMC) also aims to utilize this machine for advanced research and development tasks, anticipating a performance leap that is several times better than earlier systems.

    Vision for the Future

    Company executives view this project as a stepping stone toward a larger AI ecosystem. Jensen Huang, Nvidia’s chief executive, described AI as the driving force of “a new industrial revolution.” In contrast, Foxconn chair Young Liu emphasized the aim of linking public agencies with private enterprises through shared infrastructure. TSMC chair C.C. Wei sees the cluster as a way to accelerate semiconductor innovation.

    Foxconn plans to use the supercomputer for its internal purposes, such as enhancing smart-city services, refining driver-assistance features for its electric-vehicle platform, and improving manufacturing processes using digital-twin technology. Taiwan’s science minister, Wu Cheng-Wen, has expressed that the larger goal is to create “a smart AI island” that connects citizens, businesses, and the government through cutting-edge computing resources.

    Lastly, the supercomputer aligns with Nvidia’s broader strategy for expansion. Recently, the company announced the opening of a research and development center in Shanghai and indicated intentions to create a fleet of AI servers worth half a trillion dollars in the United States.

  • MS-C931 Mini PC with 128GB RAM: Nvidia Outperforms Intel and AMD

    MS-C931 Mini PC with 128GB RAM: Nvidia Outperforms Intel and AMD

    Key Takeaways

    1. The MSI EdgeXpert MS-C931 is marketed as an AI supercomputer, designed for local AI model processing, offering cost-effectiveness and better privacy compared to cloud services.
    2. It has compact dimensions (5.9 x 5.9 x 2 inches) and is versatile for various applications, including robotics and traffic monitoring.
    3. The device features an Nvidia Blackwell graphics card, an ARM SoC with 20 cores, and claims an AI performance of 1,000 TOPS.
    4. It can manage large language models (LLMs) with up to 200 billion parameters locally, and this capacity doubles when two units are connected.
    5. The mini PC supports up to four displays, includes WiFi 7 and Bluetooth 5.3, offers 1TB of M.2 SSD storage, and has a 10Gbit/s Ethernet port.


    We’ve talked about MSI products several times before, even if not all of them are meant for regular consumers. MSI also has offerings for businesses and professionals, and the new EdgeXpert MS-C931 fits right in. This mini PC is marketed as an AI supercomputer, which suggests it has the power to run AI models locally. This comes with several benefits. For one, it can be more cost-effective than using cloud-based AI services, and it also offers better privacy. This aspect could be vital for consulting firms that handle sensitive information from clients.

    Design and Versatility

    The MSI EdgeXpert has dimensions of 5.9 x 5.9 x 2 inches, and it can be utilized for various applications, including robotics. Robots, for instance, could leverage real-time image recognition when linked to the MS-C931 through a local network. Additionally, this mini PC could be effective for monitoring or controlling traffic. It is powered by an Nvidia Blackwell graphics card along with an ARM SoC that boasts 20 cores. The AI capability of this device is claimed to reach 1,000 TOPS, a significant figure compared to the NPUs in Intel and AMD CPUs, which typically have AI performance in the double digits. Plus, it includes 128GB LPDDR5x RAM and NVLink C2C for smooth access between the processor and GPU to the memory.

    Performance and Connectivity

    MSI states that a single unit of this mini PC can manage an LLM with up to 200 billion parameters locally. If two MSI MS-C931 units are connected, that number effectively doubles. Essentially, this system can also function like a regular mini PC without any complications. It supports up to four displays via USB Type-C, which can also be used for connecting various peripherals. It comes equipped with WiFi 7 and Bluetooth 5.3, while the M.2 SSD offers a storage capacity of 1TB. Additionally, the Ethernet port supports a bandwidth of 10Gbit/s, and this mini PC runs on a 240W PSU.

    Source:
    Link

  • European Open Web Index Pilot Grants Access to 1 Petabyte of Data

    European Open Web Index Pilot Grants Access to 1 Petabyte of Data

    Key Takeaways

    1. The Open Web Index (OWI) will launch a pilot program next month, providing access to nearly 1 petabyte of web data, with plans to expand to 5 PB and eventually 10 PB.

    2. The OWI serves as a collective digital library, allowing third-party services to search for documents, reducing Europe’s dependence on US-based search engines.

    3. The initiative aims to improve search quality and language options, promoting a non-profit, standards-based index that complies with European data protection laws.

    4. During the pilot phase, academic groups, startups, and developers can request data under research or commercial licenses, contributing to user-focused improvements.

    5. The project aligns with the European Commission’s InvestAI initiative, which seeks to boost funding for AI projects, potentially enhancing European competitiveness in search and AI technologies.


    The OpenWebSearch.eu group is set to launch the first federated Open Web Index (OWI) across Europe for external testers next month. This pilot program will provide access to nearly one petabyte of web data that has been collected, marking a significant move towards a comprehensive index planned to grow to 5 PB and eventually to 10 PB of content.

    A New Way to Search

    The OWI is not like a traditional search engine; instead, it acts as a collective digital library that allows third-party services like search portals, large language model providers, or research teams to search for and find documents. This initiative is backed by a collaboration of 14 universities, supercomputing centers, tech companies, and CERN, aiming to lessen Europe’s reliance on proprietary indexes from American companies like Google and Microsoft.

    Challenging the Status Quo

    Supporters of this project say that the current focus on advertising-driven platforms has hurt the quality of search results and restricted language options. By creating a non-profit, standards-based index in line with European regulations, the consortium hopes to promote services that abide by local data protection laws, offer results in various languages, and avoid aggressive advertising or biased results. Regulators in Brussels and London have often criticized the dominance of US tech giants for these very reasons.

    During the pilot phase, academic groups, startups, and individual developers will be able to request the dataset under a general research license or apply for a commercial license. Community manager Ursula Gmelch refers to this launch as “a first step towards true European digital sovereignty,” noting that initial feedback will help shape the index to better meet the needs of users. The team is particularly keen on enhancing vertical and argumentative search, retrieval-augmented generation, and other AI-related applications.

    Aligning with European Goals

    This timeline coincides with the InvestAI initiative from the European Commission, which aims to raise €200 billion for AI projects. An open Zoom meeting is planned for June 6, from 10 a.m. to noon CEST, where participants will be introduced to the platform and receive access credentials. If the pilot is successful, it could provide small and mid-sized European companies with the essential tools to develop competitive search and AI technologies that operate outside of the dominant US ecosystems.

    Source:
    Link

  • Dell Pro Max Plus Workstations Boosted by Qualcomm AI Technology

    Dell Pro Max Plus Workstations Boosted by Qualcomm AI Technology

    Key Takeaways

    1. Dell has launched a new Pro Max Plus laptop series designed for AI tasks, featuring the Qualcomm AI-100 Inference Card.
    2. The AI-100 Inference Card includes dual chips, 32 AI cores, and 64GB of LPDDR4x memory for high-performance AI inference.
    3. The laptops are equipped with Intel Core Ultra 7 processors, a 14-inch FHD+ touchscreen, DDR5 memory, and Arc Pro graphics.
    4. The devices are targeted at data scientists and AI engineers needing portable solutions for AI workloads.
    5. No official release date is set, but the laptops are aimed at enterprise clients and specialized developers rather than general consumers.


    Dell has recently unveiled a new version of its Pro Max Plus laptop series, specifically aimed at handling AI tasks. While the name might not be the catchiest, the technology inside these devices is more captivating than typical updates. The new Dell Pro Max Plus models will be among the first portable workstations to include a Qualcomm AI-100 Inference Card.

    Innovative Hardware for AI

    The AI-100 Inference Card, which features dual chips, is integrated into these machines to provide a secure, high-performance local solution for extensive AI inference. Dell claims the card boasts 32 AI cores and 64GB of LPDDR4x memory, which can manage models with parameters ranging from 30 billion to 109 billion. This places it firmly in the realm of enterprise-grade technology, yet it’s designed to be mobile and developer-friendly. It targets data scientists, AI engineers, and research experts who need portable solutions for testing and deployment purposes.

    Specifications and Features

    Although Dell has not released complete technical specifications for the Qualcomm model, the current Intel-based versions give a helpful point of comparison. These devices are powered by Intel Core Ultra 7 255H processors, equipped with a 14-inch FHD+ touchscreen (300 nits), DDR5 memory, Arc Pro graphics, and batteries up to 72Wh with ExpressCharge capabilities. The Qualcomm model is expected to keep most of the same design and connection features, including support for Wi-Fi 7 and Bluetooth 5.4, along with USB Type-C charging. This new setup could provide a secure way to run models locally, avoiding the need to send sensitive data to a server while still maintaining the speed and performance usually found in GPU-intensive systems.

    Target Audience and Availability

    There’s currently no official timeline for when these will be available, but this configuration is likely to attract enterprise clients or specialized developers rather than everyday consumers. Nevertheless, for those who need to prototype, test, or deploy AI workloads while traveling or in the field, the Pro Max Plus featuring Qualcomm’s inference card could be worth keeping an eye on.


  • NVLink Fusion Boosts Low-Latency Computing for Third-Party CPUs

    NVLink Fusion Boosts Low-Latency Computing for Third-Party CPUs

    Key Takeaways

    1. Introduction of NVLink Fusion: Nvidia launched NVLink Fusion, a new chip-level interface that enhances its NVLink technology for third-party CPUs and custom accelerators, announced at Computex 2025.

    2. Chiplet Technology Transition: NVLink Fusion changes connections from board-to-board to a compact chiplet setup, providing up to 14 times more bandwidth than standard PCIe while ensuring memory-semantic access.

    3. Collaborations and Partnerships: Companies like MediaTek, Qualcomm, and Fujitsu are integrating NVLink Fusion into their products, enhancing collaboration to utilize this new technology.

    4. Modular Solutions for Hyperscale Operators: Hyperscale cloud operators can create large GPU clusters with NVLink Fusion-enabled components, allowing for efficient scaling without performance drops typical of PCIe setups.

    5. Positioning Nvidia as a Key Link: By licensing its technology, Nvidia aims to be a central player in the AI hardware ecosystem, allowing competitors to use its bandwidth while still fitting into its software and networking environment.


    Nvidia has unveiled NVLink Fusion, a new chip-level interface that expands its proprietary NVLink technology beyond just its processors. This was announced during Computex 2025. The innovative silicon allows third-party CPUs and custom accelerators to utilize the same high-bandwidth, low-latency connection that currently links Nvidia GPUs in large-scale “AI factories.”

    Transition to Chiplet Technology

    NVLink Fusion shifts the connection from a board-to-board setup to a compact chiplet that designers can position alongside their own compute dies. Although it continues to utilize familiar PCIe signaling, it offers up to 14 times more bandwidth than a standard PCIe lane while maintaining memory-semantic access between devices. This enhanced fabric works alongside Nvidia’s existing Spectrum-X Ethernet and Quantum-X InfiniBand products, which manage scale-out traffic across server racks.

    Collaborations and Partnerships

    Multiple partners have already committed to this technology. MediaTek, Marvell, Alchip, Astera Labs, Cadence, and Synopsys are set to provide custom ASICs, IP blocks, or design services utilizing this new protocol. On the CPU front, Fujitsu is planning to integrate its upcoming 2 nm, 144-core Monaka processor with NVLink Fusion, while Qualcomm aims to connect the interface with its Arm-based server CPU. Both companies are targeting integration into Nvidia’s rack-scale reference systems without sacrificing direct GPU access.

    Modular Solutions for Hyperscale Operators

    Hyperscale cloud operators can now combine NVLink Fusion-enabled components with Nvidia’s own Grace CPUs and Blackwell-class GPUs, enabling them to create large GPU clusters linked by 800 Gb/s networking. This offers a modular approach to building clusters that can consist of thousands or even millions of accelerators, all without the usual performance drops seen in PCIe-only setups.

    By licensing a key part of its technology stack, Nvidia is positioning itself as the essential link for diverse AI hardware rather than just a closed-box supplier. Competitors who found it difficult to match NVLink’s impressive bandwidth can now leverage it, but they’ll need to do so within Nvidia’s extensive software and networking ecosystem.

    Source:
    Link

  • Nvidia’s Next Big AI Move: Cloud-Connected Humanoid Robots

    Nvidia’s Next Big AI Move: Cloud-Connected Humanoid Robots

    Key Takeaways

    1. Nvidia’s GPUs have driven significant growth in the AI industry, establishing a near-monopoly in AI hardware.
    2. CEO Jensen Huang envisions physical AI and robotics as the next industrial revolution, providing essential tools for robotics development.
    3. Nvidia launched Isaac, its first open-source humanoid robot operating system, allowing developers to enhance robots with trained models.
    4. The training process for robots involves creating videos from single images to teach new tasks using compressed action tokens.
    5. Nvidia supports developers with Universal Blackwell Systems powered by RTX PRO 6000 GPUs, facilitating powerful robot training capabilities.


    The AI industry has experienced huge growth over the past four years, largely due to Nvidia and its advanced GPUs that are specially made for AI tasks. Team Green started investing in AI back in the mid-2010s, but the real impact became noticeable only in recent years. With the hardware segment of AI now well established, and almost monopolized, Nvidia is exploring new opportunities that might increase its value even more. At this year’s Computex, Team Green hinted that one of these opportunities could be in physical AI and humanoid robotics.

    CEO’s Vision for the Future

    During his keynote earlier today, CEO Jensen Huang shared, “Physical AI and robotics will lead to the next industrial revolution. From AI brains for robots to simulated worlds for practice, or AI supercomputers for training foundational models, NVIDIA offers essential tools for each phase of the robotics development process.”

    Expanding Software Horizons

    With a focus on physical AI, Nvidia is also broadening its efforts in the software domain. The main element here is the company’s first open-source humanoid robot operating system named Isaac. Developers can build on this by adding trained models like the Gr00t N1.0, which was launched in March and has now been updated to version 1.5, introducing the Dreams component. Version 1.0, which featured the Mimic component, served as a foundation for training robot reasoning and actions. With the Dreams component, Nvidia is unveiling a framework that can produce large amounts of synthetic motion data (neural trajectories) that physical AI developers can use to teach robots various motor skills, including adapting to different environments.

    Training Process Explained

    The training method is similar to text-to-image and video models. Initially, the robots undergo post-training with Cosmos Predict world foundation models (WFMs). By using just one image as input, GR00T-Dreams can create videos of the robot doing new tasks in unfamiliar settings. The blueprint extracts action tokens (compressed data pieces easily handled by the robot’s neural network), which instruct the robot on how to carry out new actions.

    Nvidia’s Isaac GR00T is strongly connected to the Omniverse and Cosmos platforms to generate training data. The N1.5 update, along with models such as Isaac Sim 5.0 and Isaac Lab 2.2, will soon be downloadable from the Hugging Face repository. Innovative companies already utilizing the Isaac N models for humanoid robots include AeiRobot, Foxlink, Foxconn, Lightwheel, NEURA Robotics, Boston Dynamics, Agility Robotics, and XPENG Robotics.

    Developer Support and Hardware

    In addition, Nvidia is providing its Universal Blackwell Systems powered by RTX PRO 6000 GPUs, like the DGX Cloud infrastructure, enabling developers to access significant processing power for robot training with ease.

    Source:
    Link