Tag: AI chips

  • Nvidia Slams US Last-Minute Restrictions on AI Chip Exports

    Nvidia Slams US Last-Minute Restrictions on AI Chip Exports

    Nvidia is strongly against the forthcoming chip export limitations from the White House, characterizing them as a hurried initiative that might exceed its intended objectives. The anticipated regulations—expected to be announced imminently—aim to establish a three-tiered system to manage AI chip exports, tailored both to specific countries and companies.

    Concerns Over Policy Impact

    Ned Finkle, who leads Nvidia’s government relations, expressed his worries regarding the extensive implications of this policy. He stated that the proposed “extreme country cap policy” would negatively affect everyday computing worldwide without genuinely enhancing national security. These new restrictions would particularly influence AI accelerators, an area where Nvidia holds a dominant position.

    Proposed Access Levels

    According to the suggested rules, American semiconductors would be allocated based on different access categories. Some of the U.S.’s closest allies would receive unrestricted import permissions, while many other countries would face new limitations on overall computing capacity. These restrictions would not only target specialized AI chips but also general-purpose GPUs utilized in various devices, including gaming PCs and data centers.

    Timing and Economic Concerns

    The timing of this announcement is significant, occurring less than two weeks prior to the presidential changeover. “This last-minute policy from the Biden administration could become a legacy that will draw criticism from both the U.S. industry and the global community,” Finkle warned, cautioning that it may damage American economic interests.

    Meanwhile, Nvidia’s CEO, Jensen Huang, has expressed his willingness to work with the new administration coming in. He showed interest in meeting with Trump and even offered his assistance. Speaking at CES in Las Vegas, Huang seemed optimistic about the possibility of reduced regulations under Trump, stating, “As an industry, we want to move fast.”

    Market Implications for Nvidia

    These new regulations could significantly impact Nvidia’s position in the market, especially considering the company’s remarkable growth: its stock price nearly tripled last year, following a 239 percent increase in 2023, largely fueled by a rise in AI investments.

    Source:
    Link

  • Run:ai Joins Nvidia, Announces Open-Source Software Plans

    Run:ai Joins Nvidia, Announces Open-Source Software Plans

    AI chip giant Nvidia has finalized the purchase of Israeli AI start-up Run:ai, after getting the green light for the merger from the European Commission.

    Nvidia had initially shared its intentions to buy the start-up back in April, but the deal faced a review from the European Commission, which approved it just this month. The orchestration software firm stated that it "will keep assisting customers in maximizing their AI Infrastructure."

    Open-source Ambitions

    Run:ai has expressed its intention to make its software stack open-source "to support the community in building better AI, quicker." At present, Run:ai’s workflows are compatible only with Nvidia GPUs, but the firm is optimistic that this will evolve once the software is made available "to the whole AI ecosystem."

    Market Share and Scrutiny

    As per Nasdaq, Nvidia commands an 80% stake in the rapidly growing data center market for AI chips. However, the company is also facing intense scrutiny from lawmakers regarding potential monopoly issues.

    Earlier this month, China’s Administration for Market Regulation revealed it is conducting an investigation, and the US Commerce Department is also looking into how Nvidia’s chips made their way to China despite existing restrictions. Furthermore, the US Department of Justice is delving into the company following complaints from competitors.

    Reuters | Nvidia | Run:ai | Nasdaq

    Source: Link

  • AMD Job Cuts: Company Focuses on AI Development and Strategy

    AMD Job Cuts: Company Focuses on AI Development and Strategy

    AMD has made the decision to cut its workforce by 1,000 employees globally. This semiconductor firm, often viewed as a key competitor to Nvidia, shared an earnings report for Q3 in September that showed mixed results.

    Financial Performance Insights

    According to a report from Reuters, AMD saw its revenue in the data center sector, which includes AI chips, increase more than double in the last quarter. Additionally, the personal computer segment experienced a growth of 29%. However, the gaming division faced a significant drop, with sales falling by 69%.

    Future Projections

    The London Stock Exchange Group (LSEG) has predicted that AMD’s data center segment will grow by 98% in 2024, which greatly surpasses the expected 13%. Even with this growth, AMD still trails behind its rivals Nvidia and Intel in the AI chip market. In the first quarter of 2024, Nvidia maintained a 65% market share, while Intel held 22%, and AMD was at 11%.

    An AMD representative commented to Reuters, "In order to align our resources with our biggest growth opportunities, we are implementing several targeted measures." They also mentioned that the company is "dedicated to treating affected employees with dignity and assisting them during this transition."

    Conclusion

    In summary, while AMD is showing signs of growth in certain areas, it faces significant challenges in the competitive landscape of AI chips and has had to make tough decisions regarding its workforce. The company’s commitment to supporting its employees through these changes is notable.

    Source: Link,Link

  • India’s First AI Chips Launched by a Ride-Hailing Company

    India’s First AI Chips Launched by a Ride-Hailing Company

    Ola Electric, a company renowned in India for its electric scooters and ride-hailing services, has taken the industry by surprise with its announcement to venture into AI hardware. The company revealed its strategy to create a series of AI chips, marking a significant step towards India’s goals in the expanding field of artificial intelligence.

    AI Chip Models and Features

    The initial lineup of chips, set to be released in 2026, consists of three models: Bodhi 1, Ojas, and Sarv 1. Bodhi 1, recognized as the first AI chip to be designed and manufactured in India, focuses on AI inferencing. It is poised to be a formidable option for large language models (LLMs) and applications that need high-performance vision processing. Ola asserts that Bodhi 1 stands out in power efficiency, an essential attribute as AI systems increasingly demand more power.

    Ojas, the second chip, is tailored to meet the rising demand for edge AI solutions. It is versatile enough for a range of applications in the automotive, mobile, and IoT sectors. Ola intends to incorporate Ojas into its future electric vehicles, potentially enhancing capabilities like charging optimization and advanced driver assistance systems (ADAS).

    Sarv 1 and Market Challenges

    The third chip, Sarv 1, is a general-purpose server CPU utilizing the Arm instruction set, crafted to address the growing AI computational requirements of the data center industry.

    During Ola’s presentation, the company showcased impressive performance and power efficiency metrics for their prototype chips, making comparisons with Nvidia GPUs. However, some important details were omitted, such as the specific Nvidia GPU model used as a benchmark and the location where these chips will be manufactured.

    India’s Position in the Global AI Market

    This development highlights India’s aspiration to join the global AI competition, which is currently dominated by the US and China. With a large pool of tech talent, India is well-positioned to advance in AI. Furthermore, the ongoing restrictions on the sale of advanced technology to China by companies like Nvidia and ASML may make India an appealing alternative market for these firms.

    Nonetheless, Ola is confronted with substantial obstacles to make its bold AI chip initiative a reality. The AI hardware sector is currently ruled by well-established companies, and Ola must prove that its chips can hold their own in terms of performance, power efficiency, and cost. Additionally, the company needs to build a robust manufacturing infrastructure to produce these sophisticated silicon components.

  • TSMC’s Advanced Packaging Capacity Fully Booked by Nvidia and AMD

    TSMC’s Advanced Packaging Capacity Fully Booked by Nvidia and AMD

    TSMC, the prominent semiconductor manufacturer globally, has declared that its advanced packaging capacity has been completely reserved for the next two years. This announcement coincides with Nvidia, AMD, and Guanghuida securing TSMC's cutting-edge packaging technologies for their high-performance computing (HPC) endeavors.

    Growing Demand for AI Processors

    The emphasis on high-performance computing is driven by its crucial role in supporting artificial intelligence (AI) tasks. TSMC foresees a substantial revenue increase from AI processors, with estimates suggesting a doubling of revenue just this year. Projections indicate that over the next five years, the compound annual growth rate for AI chips will reach 50%, with AI processors anticipated to contribute more than 20% of TSMC's revenue by 2028.

    Key Players Embrace TSMC's Technologies

    Nvidia and AMD have both secured TSMC's Chip-on-Wafer-on-Substrate (CoWoS) and System-on-Integrated-Chip (SoIC) advanced packaging capacities for their respective products. Nvidia's flagship H100 chip, produced on TSMC's 4nm process, utilizes CoWoS packaging. In contrast, AMD's MI300 series, fabricated using TSMC's 5nm and 6nm processes, employs SoIC for CPU and GPU integration before incorporating CoWoS with High Bandwidth Memory (HBM).

    Guanghuida, an emerging player in the AI chip market, has also reserved TSMC's packaging capacity. Their H100 chips, powered by TSMC's 4nm process and CoWoS packaging, feature SK Hynix's HBM for improved performance. Furthermore, Guanghuida's latest Blackwell architecture AI chip, based on TSMC's advanced 4nm process, showcases upgraded HBM3e memory, doubling the computing power compared to earlier versions.

    Meeting the Escalating Demand

    The increasing demand for AI chips is being fueled by major global cloud service providers like Amazon AWS, Microsoft, Google, and Meta, all striving for dominance in the AI server sector. To tackle shortages from leading manufacturers such as Nvidia, AMD, and Guanghuida, these cloud giants are turning to TSMC to fulfill their orders, contributing to the chipmaker's positive revenue forecasts.

    To address this rising demand, TSMC is enhancing its production capacity for advanced packaging. By the year's end, CoWoS monthly production is expected to triple, reaching 45,000 to 50,000 wafers, while SoIC capacity is set to double, hitting 5,000 to 6,000 wafers. By 2025, SoIC monthly production is projected to double once more, reaching 10,000 wafers.

    The full booking of TSMC's advanced packaging capacity signifies the rapid pace of innovation in AI-driven computing, with key industry players strategically positioning themselves to leverage this burgeoning market.

  • Meta Reveals Next-Gen AI Chip “Artemis”

    Meta Reveals Next-Gen AI Chip “Artemis”

    Meta Platforms recently unveiled details about its forthcoming artificial intelligence chip dubbed Artemis, designed to meet growing processing power demands for running AI features across their social media platforms such as Facebook, Instagram and WhatsApp.

    Addressing Dependency On External Suppliers

    Artemis represents Meta’s move away from external suppliers like Nvidia for AI chip production and toward in-house chip development, in an effort to take back control over their hardware infrastructure while cutting energy consumption costs.

    Optimize Architecture For Improved Functionality

    Meta has designed Artemis with its dual objectives in mind. First, Artemis’ architecture was carefully engineered to balance computing power, memory bandwidth and capacity efficiently – thus making Artemis ideal for improving ranking and recommendation systems that form integral parts of social media platforms’ operations.

    Future Growth And Collaboration Strategies

    Meta has not shied away from its partnership with Nvidia since Artemis debuted, with CEO Mark Zuckerberg affirming their intentions of purchasing significant quantities of Nvidia H100 flagship chips throughout 2019. Meta plans on amassing over 600k AI chips by 2024 from various providers other than just Nvidia alone.

    Artemis was developed using Taiwan Semiconductor Manufacturing Co’s (TSMC) cutting-edge 5nm process and boasts an incredible threefold increase in performance when compared with Meta’s initial AI processor. Artemis is now deployed within Meta data centers worldwide supporting an array of AI applications while Meta plans on expanding Artemis to accommodate generative AI workloads in near future.

  • Microsoft & OpenAI plan $100B supercomputer Stargate AI

    Microsoft & OpenAI plan $100B supercomputer Stargate AI

    According to a recent report from The Information, Microsoft and OpenAI are in discussions for a collaborative data center project that could amount to an estimated $100 billion. This initiative aims to introduce a groundbreaking artificial intelligence supercomputer known as "Stargate" by the year 2028.

    Microsoft and OpenAI's Joint Data Center Project

    This ambitious undertaking reflects the escalating demand for AI data centers capable of managing increasingly intricate tasks, largely driven by the emergence of generative AI technology.

    Microsoft is anticipated to take on the primary financial responsibility for this venture, with projections suggesting that the project's budget will be a hundredfold greater than current data center operations.

    Key Phases of the Stargate Supercomputer Initiative

    The envisioned supercomputer, situated in the United States, will serve as the focal point of a comprehensive six-year initiative, segmented into five distinct phases. The deployment of Stargate is envisioned as the final phase, following Microsoft's ongoing development of a smaller supercomputer for OpenAI, set for unveiling in 2026.

    The acquisition of specialized AI chips is slated to be a substantial expense during the latter stages of the project, with estimates indicating a price range of $30,000 to $40,000 per chip, as highlighted by Nvidia CEO Jensen Huang.

    Financial Projections and Technological Advancements

    Microsoft's commitment to enhancing AI capabilities is exemplified through its creation of bespoke computing chips. The new data center project is designed to integrate chips from diverse suppliers, as detailed in the report.

    The estimated expenditures for the project could potentially exceed $115 billion, surpassing Microsoft's previous year's capital investment in infrastructure by a significant margin.

    Recently, OpenAI CEO Sam Altman unveiled details about the forthcoming GPT-5 model during the World Government Summit in Dubai, revealing its anticipated advancements surpassing prior iterations. GPT-5 is poised to represent a substantial leap in AI capabilities, with early demonstrations showcasing its proficiency in deciphering ancient languages. It is expected to feature functionalities such as image generation through DALL-E and video generation via Sora, with a tentative rollout projected for 2024.

  • Nvidia Reveals Two New AI Chips for China Amid Export Issues

    Nvidia Reveals Two New AI Chips for China Amid Export Issues

    Nvidia’s Strategic Move in China Amid U.S. Export Restrictions

    Nvidia, a prominent player in the tech industry, is strategically maneuvering in China by unveiling two new AI chips despite the challenges posed by U.S. export restrictions. Jensen Huang, the CEO of Nvidia, recently disclosed that these chips are now available for testing by selected customers. This move signifies Nvidia’s commitment to maintaining its foothold in the Chinese market amidst the intricate web of international trade laws.

    New Chips Tailored to Comply with Regulations

    While Huang’s announcement did not delve into the specifics of the chips or the identity of the testers, there are speculations surrounding Nvidia’s plans. Insights from sources like the SemiAnalysis newsletter suggest that Nvidia is gearing up to launch three chips – H20, L20, and L2 – tailored specifically for the Chinese market. These chips are said to incorporate Nvidia’s cutting-edge technology while being adjusted to adhere to the new U.S. regulations.

    Challenges Faced by Nvidia in China

    Nvidia has encountered challenges in the Chinese market recently, primarily due to the U.S. government’s tightened export controls. These regulatory changes have had a significant impact on Nvidia’s sales in China, once a lucrative source of revenue for the company. The implementation of these measures has led to a noticeable decline in earnings from the region.

    Forward Momentum Despite Setbacks

    Despite the obstacles, Nvidia is forging ahead with its strategic initiatives. The company has opened pre-orders for the H20 chip, positioning it as a formidable competitor against industry players like Huawei. However, there have been obstacles along the way, with delays arising from difficulties faced by server manufacturers in integrating the chip into their systems.

    Nvidia’s proactive approach in introducing new AI chips tailored for the Chinese market showcases its resilience in the face of challenging circumstances. By adapting to regulatory changes and addressing market demands, Nvidia is poised to maintain its competitive edge in China’s tech landscape.

  • Huawei Slows Down Phone Production and Shifts Focus to AI Chips

    Huawei Slows Down Phone Production and Shifts Focus to AI Chips

    Huawei is making a big change in its production plans, showing a strong interest in the growing area of artificial intelligence (AI). The tech giant has chosen to boost the production of its AI chip, the Ascend 910B, even though it means slowing down the making of its Mate 60 smartphones in at least one of its factories. This move is surprising because the Mate 60 had recently done better than Apple in terms of phone sales in China, as mentioned by the South China Morning Post. Huawei’s decision to focus more on AI chips than on smartphones highlights the company’s belief in the significant impact AI will have on the tech world.

    Increased Demand for AI Chips in China

    The Ascend 910B chip is important for developing AI models, and there’s a growing demand for it in China. This increase in interest comes as Chinese AI developers are finding it hard to get their hands on high-quality AI chips from abroad, like Nvidia’s H100, because of trade restrictions set by the United States. By concentrating on the Ascend 910B, Huawei is not only finding a way around these challenges but also setting itself up as an important player in the domestic AI market.

    Chinese Tech Companies Embrace AI

    Huawei’s new direction also shows how Chinese tech companies are eager to get involved in the AI field. Even though they’re a bit behind global leaders like OpenAI and Google, companies such as Baidu are moving forward with their own AI projects. China is also ahead in making rules for AI, requiring companies to get approval before they release AI products to the public.

    Preparing for the Future

    With this shift towards AI technology, Huawei is preparing for the future, focusing on an area that could change technology in big ways. This move shows Huawei is thinking ahead, ready to face current challenges while betting on the long-term importance of AI.