Google strikes back with its own lightweight AI model

by Yaron

Google announced Thursday that it is releasing Gemini 1.5 Flash, it’s snack-sized large language model and ChatGPT-4o mini competitor, to all users regardless of their subscription level.

The company promises “across-the-board improvements” in terms of response quality and latency, as well as “especially noticeable improvements in reasoning and image understanding.”

Google initially released Gemini 1.5 Flash in May as a lighterw-eight version of its flagship Gemini 1.5 Pro model. It’s designed to perform less resource intensive inference tasks faster and more efficiently than Pro does, much as Claude 3.5 Haiku, Llama 3.1-8B and ChatGPT-4o mini do for their respective parent models.

Flash’s context window is drastically expanding with this update, growing from a paltry 8K length to 32K (roughly 50 pages of text). Granted, that’s a 4x increase in size, but even at 32K, Gemini Flash’s context window is still just a quarter the size of GPT-4o mini’s 128K window (or, about a book’s worth).

A slide showing Google Gemini 1.5 Flash features.

Google

What’s more, Google plans to “soon add” the ability to upload text and image files, either from Google Drive or the local hard drive, direct to the context window. This feature was previously restricted to the subscription tiers.

The company also announced updates on its efforts to reduce instances of hallucinations within its models. Google plans to include “links to related content for fact-seeking prompts in Gemini,” essentially providing links to the sources it cites.

The AI will do this for both traditional search and for associated Workspace apps as well. If the AI uses information gleaned through its Gmail integration, it will provide links back to the relevant emails.

These updates are available immediately to nearly all Gemini users on both web and mobile and in 40 languages. Teens who meet the minimum age requirements needed to manage their own Google accounts will receive access next week.

TikTok lays off hundreds in favor of AI moderators while Instagram blames humans for its own issues
a person using Tiktok on their phone

ByteDance, the company behind video social media platform TikTok, has reportedly laid off hundreds of human content moderators worldwide as it transitions to an AI-first moderation scheme.

Most of the roughly 500 jobs lost were located in Malaysia, Reuters reports. Per the company, ByteDance employs more than 110,000 people in total. “We’re making these changes as part of our ongoing efforts to further strengthen our global operating model for content moderation,” a TikTok spokesperson said in a statement.

Read more

  • Computing

OpenAI uses its own models to fight election interference
chatGPT on a phone on an encyclopedia

OpenAI, the brains behind the popular ChatGPT generative AI solution, released a report saying it blocked more than 20 operations and dishonest networks worldwide in 2024 so far. The operations differed in objective, scale, and focus, and were used to create malware and write fake media accounts, fake bios, and website articles.

OpenAI confirms it has analyzed the activities it has stopped and provided key insights from its analysis. “Threat actors continue to evolve and experiment with our models, but we have not seen evidence of this leading to meaningful breakthroughs in their ability to create substantially new malware or build viral audiences,” the report says.

Read more

  • Computing

Google expands its AI search function, incorporates ads into Overviews on mobile
A woman paints while talking on her Google Pixel 7 Pro.

Google announced on Thursday that it is “taking another big leap forward” with an expansive round of AI-empowered updates for Google Search and AI Overview.
Earlier in the year, Google incorporated generative AI technology into its existing Lens app, which allows users to identify objects within a photograph and search the web for more information on them, so that the app will return an AI Overview based on what it sees rather than a list of potentially relevant websites. At the I/O conference in May, Google promised to expand that capability to video clips.
With Thursday’s update, “you can use Lens to search by taking a video, and asking questions about the moving objects that you see,” Google’s announcement reads. The company suggests that the app could be used to, for example, provide personalized information about specific fish at an aquarium simply by taking a video and asking your question.
Whether this works on more complex subjects like analyzing your favorite NFL team’s previous play or fast-moving objects like identifying makes and models of cars in traffic, remains to be seen. If you want to try the feature for yourself, it’s available globally (though only in English) through the iOS and Android Google App. Navigate to the Search Lab and enroll in the “AI Overviews and more” experiment to get access.

You won’t necessarily have to type out your question either. Lens now supports voice questions, which allows you to simply speak your query as you take a picture (or capture a video clip) rather than fumbling across your touchscreen in a dimly lit room. 
Your Lens-based shopping experience is also being updated. In addition to the links to visually similar products from retailers that Lens already provides, it will begin displaying “dramatically more helpful results,” per the announcement. Those include reviews of the specific product you’re looking at, price comparisons from across the web, and information on where to buy the item. 

Related Posts

Leave a Comment