Tag: Google AI

  • Google Unveils Powerful Gemini 2.0 Pro AI Features

    Google Unveils Powerful Gemini 2.0 Pro AI Features

    Google has rolled out access to its latest AI, the Gemini 2.0 Pro experimental model. This new AI features a massive two million token input window, the largest of any Google AI to date, allowing it to manage very large text inputs. Gemini is engineered to tackle complicated prompts with these extensive inputs. Furthermore, Gemini 2.0 Pro has the ability to browse the internet and run code, while also being capable of generating code for applications.

    Performance Compared to Other Models

    In terms of performance, Gemini 2.0 Pro surpasses previous AI models from the company across various standardized large language model benchmarks. Nevertheless, it still hasn’t reached the capabilities of humans or the top-performing AIs in every category evaluated. For instance, on the LiveBench AI LLM benchmark, the experimental scores for Gemini 2.0 Pro are only 65.13, compared to Deepseek R1’s 71.57 and OpenAI’s o3-mini which scored 75.88 in high mode.

    Human Evaluation and Security Measures

    Even so, when human evaluators assess AI based on their own prompts, Gemini 2.0 Pro stands out as one of the top two AIs globally today, according to the responses it provided on the OpenLM.ai Chatbot Arena Elo ranking. Hackers may find themselves frustrated with Gemini 2.0 Pro, as it utilized self-training methods during development to minimize the chances of producing unsafe responses.

    Subscription and Availability

    Gemini 2.0 Pro is accessible to all users of Google Gemini Advanced who subscribe for $19.99 monthly. It is also available for developers using Google AI Studio and Vertex AI. Users interested in having Gemini at their fingertips can download the Gemini app on their smartphones or buy a Google Pixel 9 Pro smartphone that comes with Gemini integrated (available for purchase on Amazon).

    Source:
    Link


     

  • Google Gemini 2.0 Launches with Experimental AI Agents

    Google Gemini 2.0 Launches with Experimental AI Agents

    Google is wrapping up 2024 with a significant announcement. On Wednesday, the tech giant from Mountain View shared a wave of updates related to AI, prominently featuring the launch of Gemini 2.0. This new language model boasts cutting-edge multimodal abilities, marking what Google describes as the start of the “agentic era,” in which virtual AI agents can complete tasks on your behalf.

    Introduction of Gemini 2.0 Flash

    At first, Google is rolling out just one model from the Gemini 2.0 lineup: Gemini 2.0 Flash experimental. This model is ultra-fast and lightweight, designed to support various input and output formats. It is capable of generating images that blend text and multilingual audio, while also being able to access Google Search, execute code, and utilize other tools seamlessly. Currently, these features are available for developers and beta testers. Despite its compact size, Gemini 2.0 Flash outshines Gemini 1.5 Pro in various metrics such as factuality, reasoning, coding, and math, all while operating at double the speed. Regular users can access the chat-optimized version of Gemini 2.0 Flash on the web today, with a mobile app version arriving soon.

    Showcasing New Experiences

    Google is also revealing several exciting applications developed with Gemini 2.0. One of these is the updated Project Astra, a virtual AI agent that was first presented in May 2024. Thanks to Gemini 2.0, it can now engage in conversations across multiple languages, utilize tools like Google Search, Lens, and Maps, recall information from previous chats, and understand language with the quickness of human dialogue. Project Astra is intended for use on smartphones and smart glasses, but is currently being tested by a select group of trusted users. If you’re interested in testing this prototype on your Android device, you can sign up for the waitlist here. Additionally, there’s an impressive demo of the Multimodal Live API, which shares similarities with Project Astra, allowing real-time interaction with a chatbot through video, voice, and screen sharing.

    Exploring Project Mariner and Jules

    Another notable project is Project Mariner, an experimental Chrome browser extension that can navigate the web and accomplish tasks for you. This extension is being tested by a limited number of users in the US and utilizes Gemini 2.0’s multimodal functions “to comprehend and reason through information displayed on your browser, including pixels and web elements like text, images, code, and forms.” Google admits that this technology is still developing and may not always work reliably. However, even in its prototype state, it is quite remarkable, as demonstrated in a YouTube video.

    Additionally, Google has introduced Jules, an AI-driven code agent that integrates directly into GitHub workflows. The company claims it can manage bug fixes and repetitive tasks, allowing you to concentrate on the actual development work you wish to accomplish.

    Much of what has been announced is currently limited to early testers and developers. Google intends to incorporate Gemini 2.0 into its various products, like Search, Workspace, Maps, and more, early next year. At that point, we’ll have a clearer picture of how these new multimodal features and enhancements can be applied in real-world scenarios. There’s still no update on the Gemini 2.0 Ultra and Pro versions.

    Source: Link

  • Google Unveils Gemini 2.0 Models for New Agentic Era

    Google Unveils Gemini 2.0 Models for New Agentic Era

    Nine months after the introduction of Gemini 1.5, Google has unveiled the next significant update for its Large Language Model (LLM), named Gemini 2.0. The first model in this new lineup, Gemini 2.0 Flash, is now available as an experimental option in Google AI Studio and Vertex AI.

    Improved Speed and Functionality

    Gemini 2.0 Flash boasts "enhanced performance at similarly fast response times" and is said to be "twice as fast" compared to 1.5 Flash. The upgraded LLM supports various modes of input, including images, text, video, and audio. Additionally, it can handle mixed media, combining pictures with text, as well as multilingual text-to-speech audio.

    New Features and APIs

    This new version allows direct access to Google Search and accommodates third-party code execution along with predefined functions. Google is also launching a Multimodal Live API for developers to utilize. A version of 2.0 Flash optimized for chat will soon be accessible on both desktop and mobile browsers, with plans for a release on the Gemini mobile app in the near future.

    Advanced Prototypes

    Google’s Project Astra, a research prototype, has been upgraded with Gemini 2.0, showing improvements in dialogue, reasoning abilities, and native integration with tools such as Google Search, Lens, and Maps. This prototype can maintain up to 10 minutes of memory during a session.

    Another research effort, Project Mariner, utilizes 2.0 to comprehend complex instructions and retrieve data from a browser screen. This includes analyzing "pixels and web elements like text, code, images and forms," and employing an experimental Chrome extension to assist users in completing tasks.

    AI Code Assistant

    The third prototype, an experimental AI code assistant named Jules, can be seamlessly integrated into GitHub workflows. It possesses reasoning and logic skills to address coding challenges and formulate solutions under the supervision of developers.

    Google also revealed that it has created AI agents "using Gemini 2.0" capable of assisting users in navigating the virtual realms of video games. These agents can analyze game dynamics based solely on on-screen actions and provide real-time suggestions on what to do next in a conversational manner.

    Source: Link

  • Samsung Galaxy Chromebook Plus: Google AI & 15.6″ OLED Display

    Samsung Galaxy Chromebook Plus: Google AI & 15.6″ OLED Display

    Google has introduced its latest Chromebook, the thinnest and lightest model in the Chromebook Plus line. This new device, named the Samsung Galaxy Chromebook Plus, is equipped with various Google AI features such as Help me read, Live Translate, and more. Additionally, it has been announced that all Chromebooks will now include Gemini, Google’s AI platform.

    Specifications of the Samsung Galaxy Chromebook Plus

    The Samsung Galaxy Chromebook Plus boasts a stylish and lightweight design, featuring a 15.6-inch OLED display with a brightness of 400 nits. It is powered by an Intel Core i3 100U processor, comes with 8GB of RAM, and offers 256GB of storage. Notably, it has a Quick Insert key for quick access to tools like text editing, adding links, and emoji searches.

    Alongside Quick Insert, the Chromebook Plus includes AI capabilities such as Help me read, which provides summaries of PDFs and articles, and Live Translate, allowing users to translate content in over 100 languages. The device also features an AI-driven Recorder app for transcriptions and improved audio and visual quality for video calls.

    New AI Features and Functionality

    With the introduction of Gemini, all Chromebooks will enable users to chat and receive assistance with tasks like trip planning and image generation. New Chromebook buyers can enjoy a complimentary three-month subscription to Google One AI Premium.

    Additional features include Welcome Recap, which gives a visual summary of recent activities, and Focus mode, designed to minimize distractions. Users can also pin essential files to their home screen for quick access.

    Battery Life and Connectivity Options

    Samsung claims that the Chromebook Plus can last up to 13 hours on a single charge. The device offers a variety of connectivity options, including a full-sized HDMI port, a micro SD card slot, two USB Type-C ports, a USB Type-A port, and a combined headphone/microphone jack. It measures 11.8mm in thickness and weighs approximately 1.17 kg.

    Pricing and Availability

    The Samsung Galaxy Chromebook Plus starts at $699 in the United States and will retail for €799 in select European markets.

    (Source: Samsung PR)


  • Logan Kilpatrick, ex-OpenAI Developer Relations Head, Joins Google to Enhance Google AI

    Logan Kilpatrick, ex-OpenAI Developer Relations Head, Joins Google to Enhance Google AI

    Logan Kilpatrick has been overseeing developer relations responsibilities at OpenAI for around eighteen months. During his tenure at OpenAI, he has been actively assisting developers in creating products utilizing the OpenAI API and ChatGPT.


    Google’s Focus on Building an Exceptional AI Developer-Centric Team

    In a strategic move, Google aims to establish itself as the leading AI developer-centric team globally. Logan Kilpatrick received an invitation from notable Google executives such as Sundar Pichai, Jeff Dean, Mat Velloso, and Josh Woodward to join their team.

    Now, Logan has officially announced his departure from OpenAI via his Twitter/X account, confirming his new role at Google. He will be spearheading product development at Google AI Studio and providing support for the Gemini API, emphasizing his commitment to excellence in this new endeavor.

    His online identity, previously known as “Logan.GPT,” has been updated to reflect his actual name, although a user humorously suggested the username “Logan Gemini” as a more fitting alternative.


    Logan’s Transition to Google: A Boost for Google AI

    Having served as the Senior Developer Advocate at OpenAI, Logan has played a pivotal role in assisting numerous developers in bringing innovative products to life. His expertise and experience are anticipated to bring significant value to Google AI’s ongoing initiatives and projects.

    Google’s pursuit of talent extends beyond Logan’s recruitment, as they actively seek skilled individuals to enhance various teams within the organization. Kilpatrick’s online outreach emphasizes the company’s commitment to fostering innovation and supporting developers worldwide.

    The community’s response to Logan’s move to Google has been largely positive, with industry leaders like Brain Solis, the Head of Global Innovation at ServiceNow, applauding the decision as a substantial win for Google’s AI division.


    Diverse Professional Background: A Wealth of Experience

    Prior to his tenure at OpenAI, Logan Kilpatrick accumulated extensive experience through diverse roles. His professional journey includes over three years as an intern at NASA, where he engaged in software data analysis research, Space Academy projects, and software engineering roles. Additionally, he spent nearly two years at Apple, contributing as a software engineer and applied machine learning engineer.

    Logan Kilpatrick, ex-OpenAI Developer Relations Head, Joins Google to Enhance Google AI