Tag: DeepSeek

  • India Develops Affordable AI Models to Compete with ChatGPT

    India Develops Affordable AI Models to Compete with ChatGPT

    Artificial Intelligence is rapidly changing the landscape, with LLMs (Large Language Models) gaining significant traction due to their wide range of uses. ChatGPT stands out as a notable AI model, alongside new innovative entrants like DeepSeek from China. However, India is now stepping up to the plate, aiming to introduce its own AI model, potentially launching within this year.

    Indian Government’s Initiative for Affordable AI

    At a recent AI conference, Ashwini Vaishnaw, who is the Union Minister of Electronics and Information Technology, announced that the Indian government is in the process of developing its foundational AI model. He emphasized that this model will offer functionalities similar to those of DeepSeek and ChatGPT, but at a much lower development cost. The minister mentioned that this new AI solution could be ready in approximately 8 to 10 months.

    Focus on Local Needs and Inclusivity

    During the event organized by the Indian AI Mission, Ashwini Vaishnaw disclosed that researchers in the country are crafting an AI ecosystem framework aimed at supporting this foundational AI model. The goal is to create an experience that caters specifically to Indian users, addressing their linguistic and contextual needs. This initiative seeks to promote inclusivity while working to remove biases found in existing models.

    Computational Strength Behind the AI Development

    The Union Minister also highlighted India’s computational capabilities, noting that the domestic AI model is being developed using a facility equipped with 18,693 GPUs. In comparison, ChatGPT was trained using around 25,000 GPUs, while DeepSeek utilized 2,000 GPUs for its development.

    Cost Comparison with Existing AI Models

    Typically, using a well-known AI model like ChatGPT might cost about $3 per hour, but the Indian AI model is expected to be priced at just Rs 100 (approximately $1.15), thanks to government subsidies. This announcement comes on the heels of UC Berkeley researchers successfully replicating DeepSeek AI for a mere $30.


  • TinyZero: Affordable DeepSeek AI Clone Developed for $30

    TinyZero: Affordable DeepSeek AI Clone Developed for $30

    While it’s quite difficult to confirm the supposed development expenses of this new AI model, even the hours put in by PhD students are likely worth much more than a mere $30. Nevertheless, this recent narrative demonstrates that the upheavals in the AI industry weren’t merely a result of exaggerated reactions. It’s worth noting that some AI models for personal use that require minimal computing power have been around for a while, even prior to the emergence of DeepSeek.

    The Power of DeepSeek

    Clearly, DeepSeek operates with a significantly larger database, whereas the researchers at the University of Berkeley zero in on “reinforcement learning.” Their software doesn’t rely on human-generated data, as the AI can only confirm its own findings.

    Self-Verification Process

    This self-checking mechanism can produce extensive chains of reasoning, but it does necessitate a certain amount of processing time. This strategy is effective for straightforward mathematical problems and programming tasks, due to the simplicity of the verification involved in these scenarios.

    The AI adjusts its methods gradually to arrive at the right answer with the least number of steps possible. Users can access this AI on GitHub, where the source code and usage examples are available.

    Open Access to Data

    The creators have also made available all the data that was used in this programming trial. Since the model doesn’t incorporate extra data, TinyZero AI stands out as a very streamlined AI model. However, it is expected to perform comparably to more advanced models in specific tasks, such as solving a mathematical puzzle game.

    Regardless of whether this AI was truly developed for just $30, once foundational technologies are made publicly accessible, there will always be individuals who can enhance or refine them. After all, today’s smartphones require only a small fraction of the computing power that early 2000s desktop PCs needed and still outperform them. In the realm of AI, this progress appears to be accelerating even more rapidly.

    Source:
    Link

  • Irish DPA Targets DeepSeek Over Data Practices Concerns

    Irish DPA Targets DeepSeek Over Data Practices Concerns

    Founded in May 2023, DeepSeek is working on a large language model that claims to be way more efficient than those from OpenAI or Meta. Now, the company is facing scrutiny from European regulators. While the European Commission hasn’t gotten involved yet, the Chinese AI firm must explain how its software operates to Irish officials. These authorities are worried about how DeepSeek manages the personal information of users in Ireland.

    Concerns from Irish Authorities

    According to a report from the Irish Data Protection Commission (DPC), cited by Reuters,
    “The Data Protection Commission (DPC) has written to DeepSeek requesting information on the data processing conducted in relation to data subjects in Ireland.”

    Currently, there’s no fixed timeline for DeepSeek to respond, and no public details about potential penalties the company could face if they fail to reply or if their response indicates non-compliance with European and Irish data laws.

    Launch of DeepSeek-R1 Chatbot

    DeepSeek recently launched its first chatbot app, based on the DeepSeek-R1 model, earlier this month. The app is free and available on both iOS and Android platforms, and it has quickly become a rival to ChatGPT. It was the most downloaded free app on the US App Store shortly after its release. One immediate effect of this launch was a significant drop in Nvidia shares, which fell by 18%.

    Open-Source Data Accessibility

    All data connected to DeepSeek-R1, including its generative AI algorithms, is open-source, allowing anyone to develop tailored solutions at a low cost. Nevertheless, it’s crucial that DeepSeek ensures its personal data handling is transparent enough to facilitate easier access to the European market.

    Source:
    Link

  • Why DeepSeek Stands Out Among AI Models

    Why DeepSeek Stands Out Among AI Models

    The AI sector has been largely led by American firms such as OpenAI, Google, and Meta for quite some time. Nevertheless, the rise of DeepSeek, a new AI startup from China, has changed the international AI scene.

    DeepSeek-R1 Model Breakthrough

    DeepSeek’s newest product, the DeepSeek-R1, is gaining attention due to its impressive performance, open-source framework, and affordable development costs. As artificial intelligence continues to play a vital role in tech advancements, it’s important to recognize how DeepSeek sets itself apart from other prominent models.

    Open-Source Advantages

    Unlike U.S. AI leaders like OpenAI, DeepSeek embraces an open-source strategy. By offering its DeepSeek-R1 model under an MIT license, it allows researchers, developers, and businesses to freely access, alter, and use the technology. In contrast, OpenAI has shifted away from its original commitment to open-source, choosing to keep its newer GPT models closed off. This open-source model promotes higher transparency, encourages cooperative enhancements, and reduces the obstacles to AI integration, making DeepSeek an appealing choice for companies and independent developers who wish to avoid being confined to proprietary systems.

  • DeepSeek App Pulled from Italian Stores Due to Privacy Issues

    DeepSeek App Pulled from Italian Stores Due to Privacy Issues

    DeepSeek, a startup from China focusing on artificial intelligence, has recently faced significant regulatory challenges in Italy. Its app has unexpectedly disappeared from both Apple’s App Store and Google Play. This action comes after Italy’s data protection authority, Garante, initiated a formal investigation into how DeepSeek manages and gathers user data. Concerns surrounding data privacy and safety have put the AI firm under a microscope, mirroring similar worries expressed in the United States and Australia.

    Italian Authority Demands Clarity on Data Usage

    Italy’s privacy regulator has granted DeepSeek and its associated companies a 20-day period to reveal essential information related to their data handling practices. Authorities are requesting specifics about the types of personal data collected, how it is sourced, its intended use, and whether the information is stored on servers located in China. Additionally, they have inquired about how DeepSeek communicates data processing practices to both registered and unregistered users, especially when information is sourced via web scraping techniques.

    Privacy Issues Amid Rapid Success

    Concerns regarding privacy have escalated following DeepSeek’s rapid ascent. The launch of its AI assistant, which rivals OpenAI’s ChatGPT, saw the app quickly rise to the top of download lists across various nations, causing unease among competitors in the US tech sector. Concurrently, US officials are evaluating possible national security threats linked to the widespread use of a Chinese AI model, with the US Navy specifically cautioning its personnel against using DeepSeek.

    Data Transparency Under Fire

    Transparency in how data is managed remains a critical point of contention. According to the company’s privacy policy, user data is kept on secure servers in China and might be shared with affiliated organizations and service providers. Despite this, Euroconsumers—a group of European consumer advocates—has raised concerns regarding the sufficiency of these notifications and questioned DeepSeek’s compliance with the European Union’s General Data Protection Regulation (GDPR).

    Italy’s examination of DeepSeek is not a new development. Earlier in 2023, the country temporarily prohibited ChatGPT due to worries about user data protection. In response to these issues, OpenAI made several adjustments to its platform, including enhanced transparency about data processing, providing users with opt-out choices, and instituting age verification measures aimed at protecting children under 13. These modifications ultimately led to the reinstatement of the chatbot.

    Future Implications for DeepSeek

    As DeepSeek continues to expand its presence worldwide, the regulatory hurdles it faces are intensifying. The company is required to provide answers to the Italian regulator by February 17, a deadline that could significantly impact its future operations in the European market. Should authorities determine that privacy laws have been violated, DeepSeek may encounter severe penalties or operational restrictions, potentially setting a precedent for the examination of AI products created outside Western jurisdictions.

  • Alibaba’s Qwen 2.5 Max AI Outperforms Deepseek in Competition

    Alibaba’s Qwen 2.5 Max AI Outperforms Deepseek in Competition

    Chinese tech giant Alibaba Group Holding Ltd. has rolled out its latest AI model, Qwen 2.5 Max. The company claims this model outshines Deepseek v3—this statement comes just weeks after Deepseek was launched on January 10th, an event many were eagerly awaiting.

    Performance Claims

    Alibaba asserts that Qwen 2.5 Max exceeds the performance of other leading AI models, including those from Deepseek, OpenAI, and Meta. The model has shown remarkable results in several benchmarks, such as Arena-Hard, LiveBench, LiveCodeBench, MMLU, and GPQA-Diamond. Its performance in MMLU and LiveCodeBench has even set new industry benchmarks, showcasing its advanced capabilities.

    Market Reactions

    The introduction of Deepseek sent ripples through Silicon Valley, leading to a drop in tech stock prices and urging competitors to enhance their AI technologies. In a quick response to Deepseek’s impactful launch, ByteDance declared that it had made advancements to its own AI model, claiming it now outperforms OpenAI’s o1 in AIME benchmarks.

    Strategic Timing

    The launch of Qwen 2.5 Max seems to be timed with care, likely a reaction to the rising competition faced by China’s tech sector from international players. This announcement was made on January 29th, 2025, coinciding with the Lunar New Year, a significant holiday when many businesses in China close for a break. By unveiling this model early, Alibaba is showing its resolve to lead in AI innovation, even as competition heats up due to Deepseek’s entry.

    As the field of AI keeps advancing rapidly, Alibaba’s new achievement marks another important milestone in the global AI competition.

    Source:
    Link


  • Radeon RX 7900 XTX Beats RTX 4090 & 4080 Super in DeepSeek AI

    Radeon RX 7900 XTX Beats RTX 4090 & 4080 Super in DeepSeek AI

    Chinese LLM DeepSeek caused a significant disturbance in the US tech industry, leading to a loss of trillions of dollars from the stock market. Even though it was developed using somewhat older Nvidia hardware, it surprisingly performs quite well on AMD’s consumer product: the Radeon RX 7900 XTX. David McAfee, who oversees AMD’s Radeon division, shared some benchmark results on X.

    Performance Insights

    The Radeon RX 7900 XTX shows performance differences based on the model and the number of parameters being used. It can be up to 13% faster at 7 billion parameters and about 2% faster at 14 billion parameters. Beyond that point, the RDNA 3 flagship struggles, ultimately falling short compared to the RTX 4090 with 32 billion parameters. AMD even made a comparison with the GeForce RTX 4080 Super, where the 7900 XTX boasts a 34% advantage in performance.

    Running DeepSeek Locally

    AMD has also shared comprehensive guidelines on how to operate DeepSeek on your own computer. However, it’s important to note that the Radeon RX 7900 XTX has a limit of 32 billion parameters. On the other hand, the Strix Halo Ryzen AI Max 395 Plus, equipped with 128 GB of RAM, can handle up to 72 billion parameters. Additionally, for those who are willing to spend around $6,000, Matthew Carrigan has discovered a method to run the entire model locally on a system with dual AMD Epyc CPUs and 768 GB of RAM.

    Source:
    Link


  • China’s DeepSeek: A Major Challenge to OpenAI’s ChatGPT

    China’s DeepSeek: A Major Challenge to OpenAI’s ChatGPT

    From November 2023, DeepSeek, a Chinese firm, has started to roll out its AI models as open-source. With the MIT license in place, anyone can utilize and modify the model for personal use. This openness promotes transparency and flexibility in how these models can be applied.

    Collaborative Development

    Moreover, the open-source nature fosters teamwork in development and helps save costs. Users have the ability to inspect the code, allowing them to comprehend the model’s operations. They can tailor the model to meet their unique needs and employ it across various scenarios. By embracing open-source, DeepSeek contributes to innovation and competition within the AI landscape.

    Company Background

    DeepSeek is a spin-off from Fire-Flyer, the deep-learning division of the Chinese hedge fund, High-Flyer. The primary aim was to enhance the understanding, interpretation, and prediction of financial data within the stock market. Since its establishment in 2023, DeepSeek has shifted its focus solely onto LLMs, which are AI models that can generate text.

    Major Breakthroughs

    The company appears to have made significant advances with the latest additions to the DeepSeek AI lineup. Based on popular AI benchmarks, DeepSeek-V3, DeepSeek-R1, and DeepSeek-R1-Zero frequently surpass rivals from Meta, OpenAI, and Google in their specific areas. Additionally, these services are notably cheaper than ChatGPT.

    Impact on Pricing

    This competitive pricing tactic could influence pricing trends across the AI market, making sophisticated AI technologies more accessible to a broader audience. DeepSeek is able to maintain these lower costs by investing much less in training its AI models compared to others. This is primarily achieved through more streamlined training processes and extensive automation.

    Efficiency in Reasoning Models

    Conversely, DeepSeek-R1 and DeepSeek-R1-Zero function as reasoning models. They begin by formulating a strategy to answer inquiries before proceeding in smaller increments. This method enhances result accuracy while requiring less computational power. Nevertheless, it does increase the demand for storage space.

    Accessibility of Models

    As an open-source AI, DeepSeek can operate directly on users’ computers. Users can access the necessary application data without cost, as the models are freely downloadable from Hugging Face. Tools like LM Studio simplify the process by automatically fetching and installing the full application code.

    Data Security and Privacy

    This setup ensures data privacy and security, as prompts, information, and responses remain on the user’s device. Furthermore, the model can function offline. While high-end hardware isn’t essential, ample memory and storage are necessary. For example, DeepSeek-R1-Distill-Qwen-32B needs approximately 20GB of disk space.

    Language Capabilities

    Per DeepSeek V3, the AI is capable of handling multiple languages, including Chinese and English, as well as German, French, and Spanish. In brief interactions, the various languages yielded satisfactory replies.

    Concerns Regarding Censorship

    However, there are lingering concerns about censorship in China. DeepSeek-R1 incorporates restrictions on certain politically sensitive subjects. Users attempting to inquire about specific historical events may receive no response or a “revised” reply. For instance, asking about the events at Tiananmen Square on June 3rd and 4th, 1989 may not yield clear information.

    Censorship in AI Models

    That said, DeepSeek R1 does acknowledge the student protests and a military operation. Yet, other AI systems also limit their responses to political inquiries. Google’s Gemini, for instance, outright avoids addressing questions that may pertain to politics. Thus, (self-imposed) censorship is a common trait found in various AI models.

    Source:
    Link

  • China’s DeepSeek AI Assistant Becomes Top Free iPhone App

    China’s DeepSeek AI Assistant Becomes Top Free iPhone App

    DeepSeek, a fresh and groundbreaking AI from a startup in China, is showing incredible momentum. We previously discussed how DeepSeek AI is taking on OpenAI’s stronghold with ChatGPT. In a very brief period, the DeepSeek AI assistant app has risen to the number one spot as the best free app on iPhones.

    DeepSeek AI Assistant Dominates Apple’s App Store

    At this moment, the Chinese AI assistant from DeepSeek holds the title of the top-rated free app on Apple’s App Store in the United States. It has surpassed competitors like OpenAI’s ChatGPT. For those who may not know, DeepSeek is creating a buzz for being an open-source platform, built upon the DeepSeek V3 model. The AI reportedly uses much less computing power compared to its rivals.

    Overview of DeepSeek AI Assistant

    While still a matter of debate, the creators of DeepSeek assert that it was developed for less than 6 million US Dollars. Even with a relatively low-cost development process, it claims to deliver performance on par with Claude 3.5, GPT-4o, and others. Like many other AI platforms, DeepSeek provides features similar to ChatGPT, which can assist content creators and support research efforts. It is readily available as an app, an API, and online, making it easy for everyone to access.

    The initial DeepSeek-R1 was launched under an MIT license, allowing for commercial usage without any limitations. DeepSeek’s open-source strategy also challenges the numerous closed-source models created by large tech firms. This move toward increased transparency and accessibility enables a broader array of individuals and organizations to engage in its development and take advantage of its potential.

  • Chinese AI Startup DeepSeek Challenges OpenAI’s Dominance

    Chinese AI Startup DeepSeek Challenges OpenAI’s Dominance

    A new contender has emerged in the tech arena—DeepSeek, a Chinese AI startup, is making waves in Silicon Valley with its budget-friendly language model, DeepSeek-R1, which competes with OpenAI’s ChatGPT. Despite facing restrictions from the US on advanced AI chips, this startup has made significant strides by implementing creative strategies that emphasize both efficiency and performance. This progress is transforming the AI landscape—keep reading for more insights.

    DeepSeek’s Innovative Models

    In contrast to numerous Western AI firms that thrive on amassing extensive computing power, DeepSeek has adopted a unique strategy. The company has concentrated on enhancing software and algorithms to boost efficiency, especially given the constraints imposed by US export regulations on advanced chips. DeepSeek presents two sophisticated AI models: DeepSeek-V3, which is versatile for various applications, and DeepSeek-R1, an economical substitute for ChatGPT.

    Versatile Applications

    DeepSeek-V3 is a cutting-edge AI language model that caters to a wide array of applications, from natural language processing to customer service, education, and healthcare. Its design is particularly attuned to the Chinese language and its cultural nuances, while also accommodating global use cases. The model prioritizes high performance and affordability, positioning it as a flexible asset for multiple industries, especially within the Chinese market, yet adaptable for international use as well.

    Competitive Edge

    On the other hand, DeepSeek-R1 stands out as another model that offers performance on par with OpenAI’s ChatGPT but at a much lower price point. Even with the hurdles posed by US restrictions on advanced AI chips, DeepSeek-R1 continues to deliver high-quality outcomes through its focus on efficiency and innovative methodologies. The model aims to be a budget-friendly option compared to other AI models like ChatGPT, establishing DeepSeek as a formidable player in the global AI scene. By tackling resource challenges head-on, DeepSeek-R1 reflects the company’s dedication to innovation and scalable performance.

    Liang Wenfeng, the founder of DeepSeek and a former quant hedge fund manager, has brought together a team of enthusiastic young researchers from leading Chinese universities. He provides them the necessary resources and autonomy to pursue unconventional ideas. This nurturing environment has facilitated the creation of groundbreaking techniques such as Multi-head Latent Attention (MLA) and Mixture-of-Experts, which dramatically lower the computational demands for training their models.