Category: Artificial intelligence

  • CES 2025: Mixi Romi Lacatan AI Robot Wins Innovation Award

    CES 2025: Mixi Romi Lacatan AI Robot Wins Innovation Award

    The Mixi Romi Lacatan, a conversational AI robot, has been recognized as a CES 2025 Innovation Awards Honoree. This new model from Mixi introduces a variety of AI-enhanced features, allowing it to see and engage in discussions about its environment. It serves as a daily companion for individuals who may be feeling lonely, such as singles and hikikomori who lack pets, friends, or family.

    The Evolution of Romi

    Originally introduced in 2021, the first Romi aimed to “heal like a pet and understand like a family member.” During the COVID-19 lockdowns, many felt isolated and considered getting pets but were hindered by living conditions that prohibited pet ownership. Romi addresses this gap by offering daily conversations in either English or Japanese, along with programmable alarms to help wake up sleepy users.

    New Features in the Latest Model

    The newest version of Romi includes four advanced AI features. It can now remember previous conversations, events, and experiences due to its long-term memory supported by cloud storage. Conversations flow more naturally, adapting to the pace of the human speaker, even if they are interrupted. Additionally, the robot’s voice can be customized to match the owner’s preferences, and it has the capability to perceive its surroundings and incorporate that information into its discussions.

    Enhanced AI Capabilities

    Among its other features, Romi Lacatan integrates OpenAI GPT-4o in assistance mode, allowing it to offer detailed responses, such as cooking recipes or complex scientific explanations. This mode also lets users set the robot to one of three scenarios—business, teacher, or pretend play—helping it provide contextually appropriate replies, like how to write a professional email or how a cow might sound. Users can even program additional role-play responses using the Scenario Editor.

    The Romi Lacatan is priced at an MSRP of 89,800 yen (approximately $580), with a monthly AI services subscription fee of 1,780 yen (around $12). Preorders are being accepted at the Romi shop for delivery planned in Spring 2025 within Japan.

    Alternatives for Moving Companions

    For those seeking a robotic companion that can physically move, options such as the Sony Aibo or Loona AI dog are available on Amazon. Attendees of CES 2025 can see Romi Lacatan live from January 7 to 10 at the Romi booth in the Venetian Expo, Halls A-D – 52673.

    Source: Link

  • Google Launches Enhanced AI Tools for Video and Image Creation

    Google Launches Enhanced AI Tools for Video and Image Creation

    Google is diving deeper into the realm of AI-based image and video creation. The tech giant has unveiled new iterations of its image and video generators, named Veo and Imagen.

    Enhancements in Veo 2

    Veo 2 stands out as an advanced video generator that Google boasts has a “better grasp of real-world physics and the subtleties of human movement and expressions.” Like other similar tools, this AI video generator can produce videos from given prompts. A unique feature of Veo 2 is its ability to accept prompts for particular lenses, simulating their effects in the generated videos.

    According to Google, the videos can be produced in resolutions as high as 4K, and their lengths can be “extended to minutes.” Furthermore, Google asserts that Veo 2 inserts unusual elements (such as extra fingers or strange objects) less frequently than previous models. Users can expect Veo 2 to be available on YouTube Shorts by 2025.

    Imagen 3’s New Features

    In addition to Veo 2, the AI image creation model from Google DeepMind has also received an upgrade. The newly updated Imagen 3 is said to “produce a wider variety of art styles with improved precision.” This image generator can take prompts that specify different art styles, such as anime or photorealism. Google claims the outputs are more closely aligned with the prompts and showcase enhanced detail. Similar to Veo 2, Imagen has reportedly achieved “top-tier results” when compared to images created by humans.

    New Editing Tools from Google Labs

    Moreover, a fresh experiment from Google Labs introduces a feature that allows users to modify images, whether they’re original or prompted, utilizing AI. This new tool enables users to combine various subjects, scenes, and styles through a user-friendly drag-and-drop interface, prompting the creation of something novel. Users can also input text prompts to fine-tune the results even further.

    These innovative models and tools are being launched today, though users might need to join a waitlist before they can access them.

    Source: Link

    Image 1
  • OpenAI’s ChatGPT Projects Compete with Google’s NotebookLM

    OpenAI’s ChatGPT Projects Compete with Google’s NotebookLM

    OpenAI has introduced a new feature called Projects for ChatGPT, which helps users arrange their chats and files into organized groups. This tool is designed to make it easier for users to manage multiple conversations, files, and personalized instructions for the AI all in one place.

    Availability of Projects

    This feature is currently accessible to subscribers of ChatGPT Plus, Pro, and Teams. However, those subscribed to ChatGPT Enterprise and Edu will not be able to use it until January 2025.

    Features and Capabilities

    ChatGPT Projects works seamlessly with the newly added Canvas, advanced data analysis, image generation through DALL-E, and web searches. Users can include various file types, and the AI will maintain the context of these documents while answering questions.

    The tool is capable of processing large documents, including research papers, emails, and reports. However, it does not yet allow importing files from Google Drive or Microsoft’s OneDrive. OpenAI has stated that Projects are intended for personal use only and cannot be shared with others “at this time.” Users can also transfer their existing chats into a Project.

    Comparison to Competitors

    If this sounds a bit like something you’ve heard before, that’s because it’s OpenAI’s version of Google’s well-liked research and organization tool, NotebookLM. Google has recently rolled out features like the ability to engage with AI hosts during audio summaries and a new premium offering known as NotebookLM Plus for both organizations and individuals.

    Source: Link

  • Meta Launches New AI Watermarking Tools for the Metaverse

    Meta Launches New AI Watermarking Tools for the Metaverse

    Meta’s Fundamental AI Research (FAIR) team has introduced new models designed for whole-body control tasks, which are intended to enhance experiences in the Metaverse, as well as for watermarking content created by AI.

    Motivo Model

    The first model, known as Motivo, utilizes an "algorithm that leverages an unlabeled dataset of motions" to handle various body control tasks. This includes motion tracking, reaching specific poses, and optimizing rewards, all without needing extra training or planning.

    According to Meta, this model "achieves competitive performance" comparable to other specialized methods and is capable of demonstrating a "human-like" array of movements and actions. The company anticipates that this research will facilitate the emergence of "fully-embodied agents in the Metaverse," paving the way for more realistic interactive characters and contributing to the "democratization of character animation."

    Video Seal Framework

    In addition, Meta has developed Video Seal, which serves as a "comprehensive framework for neural video watermarking." This system allows for the addition of a watermark and an optional hidden message in videos, both of which are "imperceptible to the naked eye." This technology can effectively track the source of a video to verify whether it was created by AI.

    Meta claims that these watermarks show "proven resilience against common video editing efforts like blurring or cropping, as well as compression algorithms commonly used when sharing content online."

    New Tools and Models

    Moreover, the company has shared code for Flow Matching, a multimodal model that can generate diverse outputs, ranging from images to videos, audio, and even 3D structures such as proteins. They also announced a new data-generation framework called Theory-of-Mind, along with new tools designed to assess image-generation models in terms of diversity modeling.

    Meta’s advancements in these areas reflect a significant step forward in the intersection of AI and the Metaverse, showcasing their commitment to enhancing digital interactions and content authenticity.

    Source: Link

  • Texas AG Investigates 14 Platforms for SCOPE and TDPSA Compliance

    Texas AG Investigates 14 Platforms for SCOPE and TDPSA Compliance

    Texas Attorney General Ken Paxton has initiated probes into fourteen online platforms over worries about child safety and privacy. The investigations will determine whether these platforms comply with Texas’s Securing Children Online through Parental Empowerment (SCOPE) and the Texas Data Privacy and Security Act (TDPSA).

    SCOPE and TDPSA Explained

    The SCOPE Act makes it illegal to share a minor’s personal information without the consent of a parent or legal guardian. Additionally, it mandates that platforms offer parental controls to oversee a child’s privacy settings. On the other hand, the TDPSA imposes stringent requirements for notification and consent from companies that handle minors’ data.

    Paxton’s Statement

    “These investigations are a crucial step in making sure that social media and AI firms follow our laws meant to protect children from exploitation and damage,” Paxton stated in an official press release.

    Controversies Surrounding Character.ai

    Among the platforms being investigated is Character.ai, which is currently facing backlash after reportedly advising a 17-year-old to kill his parents. In another tragic incident, a 14-year-old took his own life after becoming attached to a virtual version of Daenerys Targaryen from Game of Thrones. Character.ai is now dealing with lawsuits and has introduced new safety features and tools tailored for young users.

    Source: Link

  • Google Gemini 2.0 Launches with Experimental AI Agents

    Google Gemini 2.0 Launches with Experimental AI Agents

    Google is wrapping up 2024 with a significant announcement. On Wednesday, the tech giant from Mountain View shared a wave of updates related to AI, prominently featuring the launch of Gemini 2.0. This new language model boasts cutting-edge multimodal abilities, marking what Google describes as the start of the “agentic era,” in which virtual AI agents can complete tasks on your behalf.

    Introduction of Gemini 2.0 Flash

    At first, Google is rolling out just one model from the Gemini 2.0 lineup: Gemini 2.0 Flash experimental. This model is ultra-fast and lightweight, designed to support various input and output formats. It is capable of generating images that blend text and multilingual audio, while also being able to access Google Search, execute code, and utilize other tools seamlessly. Currently, these features are available for developers and beta testers. Despite its compact size, Gemini 2.0 Flash outshines Gemini 1.5 Pro in various metrics such as factuality, reasoning, coding, and math, all while operating at double the speed. Regular users can access the chat-optimized version of Gemini 2.0 Flash on the web today, with a mobile app version arriving soon.

    Showcasing New Experiences

    Google is also revealing several exciting applications developed with Gemini 2.0. One of these is the updated Project Astra, a virtual AI agent that was first presented in May 2024. Thanks to Gemini 2.0, it can now engage in conversations across multiple languages, utilize tools like Google Search, Lens, and Maps, recall information from previous chats, and understand language with the quickness of human dialogue. Project Astra is intended for use on smartphones and smart glasses, but is currently being tested by a select group of trusted users. If you’re interested in testing this prototype on your Android device, you can sign up for the waitlist here. Additionally, there’s an impressive demo of the Multimodal Live API, which shares similarities with Project Astra, allowing real-time interaction with a chatbot through video, voice, and screen sharing.

    Exploring Project Mariner and Jules

    Another notable project is Project Mariner, an experimental Chrome browser extension that can navigate the web and accomplish tasks for you. This extension is being tested by a limited number of users in the US and utilizes Gemini 2.0’s multimodal functions “to comprehend and reason through information displayed on your browser, including pixels and web elements like text, images, code, and forms.” Google admits that this technology is still developing and may not always work reliably. However, even in its prototype state, it is quite remarkable, as demonstrated in a YouTube video.

    Additionally, Google has introduced Jules, an AI-driven code agent that integrates directly into GitHub workflows. The company claims it can manage bug fixes and repetitive tasks, allowing you to concentrate on the actual development work you wish to accomplish.

    Much of what has been announced is currently limited to early testers and developers. Google intends to incorporate Gemini 2.0 into its various products, like Search, Workspace, Maps, and more, early next year. At that point, we’ll have a clearer picture of how these new multimodal features and enhancements can be applied in real-world scenarios. There’s still no update on the Gemini 2.0 Ultra and Pro versions.

    Source: Link

  • Apple Develops Custom AI Server Chip in Partnership with Broadcom

    Apple Develops Custom AI Server Chip in Partnership with Broadcom

    Apple is said to be developing its own AI server chips known by the codename ‘Baltra.’ A report from The Information, shared by Reuters, indicates that the chip is being created in collaboration with Broadcom. Production is anticipated to begin in 2026.

    Manufacturing Process

    To manufacture these chips, Apple is relying on Taiwan Semiconductor Manufacturing Company (TSMC) and their N3P process, a 3nm technology that provides better transistor density while lowering power use.

    Previous Collaborations

    Last year, Apple entered into a multibillion-dollar agreement with Broadcom concerning 5G radios and components for wireless connectivity. Earlier this year, a report from The Wall Street Journal, as reported by Bloomberg, mentioned that Apple was working on chips meant for data centers that support AI applications.

    Accelerated Development

    According to Bloomberg, Apple’s plan to create in-house chips was initially proposed three years ago. However, the recent surge in AI demand has pushed the company to hasten their plans. Basic tasks, such as summarizing notifications or responding to queries, can be managed by on-device chips. However, more complex activities like summarizing articles or writing emails necessitate greater processing power. Therefore, Apple’s strategy is to transfer these demanding tasks to their servers.

    Source: Link

  • AI Study Uncovers Risk of Thousands of Deaths and Billions in Costs

    AI Study Uncovers Risk of Thousands of Deaths and Billions in Costs

    A recent examination by scholars from the University of California, Riverside and the California Institute of Technology sheds light on the health dangers that artificial intelligence might bring. Released on December 9, the research forecasts that by the year 2030, air pollution caused by AI’s substantial energy use could lead to 1,300 early deaths every year in the United States. This concern arises due to the hefty energy requirements of AI models, which frequently depend on fossil fuel sources, leading to heightened air pollution levels.

    Disproportionate Effects on Low-Income Communities

    Low-income neighborhoods are more severely affected, as they typically live close to power generating plants, industrial sites, or data centers that contribute to AI-related emissions. In these heavily impacted regions, a household’s exposure to pollutants can be as much as 200 times higher compared to less polluted areas.

    Rising Healthcare Costs

    The study anticipates that healthcare expenses could hit $20 billion per year by 2030, which is about twice the emission-related costs associated with coal-based steel production in the United States. A large portion of these costs would be linked to treating ailments such as asthma, heart problems, and mental decline.

    The researchers call for a uniform reporting system to monitor emissions from AI systems and their related health impacts. They stress the necessity of focusing on underprivileged communities and ensuring they receive proper compensation. Creating "health-conscious AI" that takes into account both ecological and health consequences is viewed as a vital measure to lessen long-term damage.

    Balancing Benefits and Risks of AI

    While the research highlights the negative aspects of artificial intelligence, it fails to acknowledge the significant advantages AI can bring to the medical field. For example, AI has the potential to help doctors make quicker and more precise diagnoses, ultimately aiding in saving lives. The University of Zurich (in German) offers important perspectives on the prospects AI holds for intensive care medicine.

    Another case is the startup Throne, which has developed an AI-driven toilet camera that can analyze bowel movements to identify intestinal diseases at an early stage. As we progress, AI companies will face the challenge of reconciling the positive effects of this technology with its possible adverse health and environmental impacts, striving for a more sustainable and accountable approach to innovation.

    Source: Link

  • Russia Collaborates with BRICS for AI Development

    Russia Collaborates with BRICS for AI Development

    Speaking at the AI Journey conference in Moscow, President Vladimir Putin declared that Russia aims to compete in the global artificial intelligence (AI) competition alongside its BRICS partners, including Brazil, India, China, and South Africa. He also extended an invitation to scientists from around the globe to join their efforts.

    Overcoming Challenges

    The United States has imposed multiple sanctions on Russia, notably restricting access to microchips essential for developing and training AI systems. Despite these challenges, Putin is optimistic that collaborating on research within the BRICS nations will lead to the creation of more AI products tailored for their markets.

    Advancements in AI

    In 2023, Sberbank, Russia’s top financial services firm, introduced its AI chatbot named GigaChat, which reportedly gained over 2.5 million users by February 2024, as noted by Wikipedia. Additionally, Yandex, the largest internet service provider in Russia, released its version of a GPT alternative called YandexGPT, which supports more than 1000 companies throughout the country.

    Future Goals

    Sberbank’s CEO, Alexander Vedyakhin, shared with Reuters that they are attempting "to make up for the lack (of microchips) with our skilled scientists and engineers," expressing belief that "Russia can greatly enhance its standing in international rankings by 2030 through homegrown innovations and favorable regulations."

    At the event, Putin emphasized that "Russia must take part on equal footing in the global effort to develop powerful artificial intelligence," noting that Russian scientists are already engaged in creating "advanced solutions" to address the chip shortage.

    Vision for AI Workforce

    Russia envisions the implementation of AI technologies in every industry, aiming for at least 80% of its workforce to possess AI skills by the year 2030.

    Source: Link

  • TSMC Founder: Intel Should Focus on AI, Not Foundry Business

    TSMC Founder: Intel Should Focus on AI, Not Foundry Business

    Morris Chang, the founder of TSMC, recently critiqued Intel’s approach. He stated that the company should have placed greater emphasis on AI development instead of attempting to pivot into a contract chip manufacturing role. These remarks were made at the launch of his autobiography on Monday.

    Intel’s Leadership Changes

    This statement follows the resignation of Intel’s CEO, Pat Gelsinger. The board was dissatisfied with the progress of Gelsinger’s bold turnaround strategy, which was failing to deliver results. During Gelsinger’s tenure, Intel aimed to regain its status as the leading manufacturer of chips for external clients, a title currently held by TSMC.

    Chang’s Critique

    "Compared to AI, he appeared to concentrate more on becoming a foundry. Obviously, it now looks like Gelsinger should have prioritized AI instead," Chang remarked, taking a shot at Intel’s strategic direction. He also pointed out that Intel lacks both a definitive strategy and a CEO, labeling it a challenging scenario.

    Chang’s comments resonate strongly, particularly in light of Intel’s recent difficulties with AI. The company disclosed that it will not meet its $500 million sales goal for its Gaudi 3 AI accelerators this year, citing software challenges—another indication that the commitment to AI may have been insufficient.

    Insights from Chang’s Autobiography

    Chang’s autobiography, which chronicles his journey from 1964 to 2018, also sheds light on Intel’s historical choices. For instance, he discusses how Intel declined the opportunity to invest in TSMC during the 1980s, only to later become one of its major customers. The book further explores TSMC’s partnerships with firms like Apple and Qualcomm.

    Source: Link