Thursday, December 19, 2024

Opera becomes the first major browser with built-in access to local AI models

Related stories

ePlus Launches Secure GenAI Accelerator

ePlus inc. announced its Secure GenAI Accelerator offering. Part of...

Lockheed Martin Launches Astris AI to Enable Secure AI Solution

Lockheed Martin has announced the formation of Astris AI,...

Boomi Expands Data Management with Rivery Acquisition

Boomi™, the intelligent integration and automation leader, announced a...

Tray.ai Launches Merlin Agent Builder to Break the Traps of Custom Code and SaaS Agents

Tray.ai, innovator of the AI-ready composable integration platform, announced...

Aily Labs and Mila Partner to Advance AI Agents and Decision Intelligence

Aily Labs GmbH, pioneer of an AI-powered decision intelligence...
spot_imgspot_img

Browser innovator Opera announced that it’s adding experimental support for 150 local LLM (Large Language Model) variants from approximately 50 families of models to its Opera One browser in developer stream. This step marks the first time local LLMs can be easily accessed and managed from a major browser through a built-in feature. The local AI models are a complimentary addition to Opera’s online Aria AI service. Among the supported local LLMs are:

  • Llama from Meta
  • Vicuna
  • Gemma from Google
  • Mixtral from Mistral AI
  • And many families more

Using Local Large Language Models means users’ data is kept locally, on their device, allowing them to use generative AI without the need to send information to a server.  Opera is testing this new set of local LLMs in the developer stream of Opera One as part of its new AI Feature Drops Program, which allows early adopters to test early, often experimental versions of the browser’s AI feature set.

Also Read: OctoAI Unveils Industry-First Generative AI Production Stack For The Enterprise

As of today, the Opera One Developer users are getting the opportunity to select the model they want to process their input with. To test the models, they have to upgrade to the newest version of Opera Developer and follow several steps to activate the new feature. Choosing a local LLM will then download it to their machine. The local LLM, which typically requires 2-10 GB of local storage space per variant, will then be used instead of Aria, Opera’s native browser AI, until a user starts a new chat with the AI or switches Aria back on.

“Introducing Local LLMs in this way allows Opera to start exploring ways of building experiences and knowhow within the fast-emerging local AI space,” said Krystian Kolondra, EVP Browsers and Gaming at Opera.

Track record or innovation in the AI space

In early 2023, Opera presented Opera One, its AI-centric flagship browser based on modular design principles and a new browser architecture with a multithreaded compositor that allows for a smoother than ever processing of UX elements. Opera One features the Aria browser AI, which can be accessed via the browser sidebar of the browser command line. Aria is also available in the gamer-centric Opera GX, as well as in Opera browser on iOS and Android.

SOURCE: PRNewswire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img