Tuesday, November 5, 2024

Opera becomes the first major browser with built-in access to local AI models

Related stories

Crayon Joins AWS Generative AI Partner Innovation Alliance

Crayon announced it will work with Amazon Web Services...

Sikich announced the appointment of Ray Beste as Principal AI Strategist

Sikich, a Chicago-based leading global technology-enabled professional services company,...

Wondershare Unveils SelfyzAI 3.0: New AI Features Enhance Image Editing Experience

Wondershare proudly launched SelfyzAI 3.0, the latest version of...

Dan Muscatello Joins OneSix as Chief Revenue Officer

OneSix, a leading data and artificial intelligence (AI) consultancy...
spot_imgspot_img

Browser innovator Opera announced that it’s adding experimental support for 150 local LLM (Large Language Model) variants from approximately 50 families of models to its Opera One browser in developer stream. This step marks the first time local LLMs can be easily accessed and managed from a major browser through a built-in feature. The local AI models are a complimentary addition to Opera’s online Aria AI service. Among the supported local LLMs are:

  • Llama from Meta
  • Vicuna
  • Gemma from Google
  • Mixtral from Mistral AI
  • And many families more

Using Local Large Language Models means users’ data is kept locally, on their device, allowing them to use generative AI without the need to send information to a server.  Opera is testing this new set of local LLMs in the developer stream of Opera One as part of its new AI Feature Drops Program, which allows early adopters to test early, often experimental versions of the browser’s AI feature set.

Also Read: OctoAI Unveils Industry-First Generative AI Production Stack For The Enterprise

As of today, the Opera One Developer users are getting the opportunity to select the model they want to process their input with. To test the models, they have to upgrade to the newest version of Opera Developer and follow several steps to activate the new feature. Choosing a local LLM will then download it to their machine. The local LLM, which typically requires 2-10 GB of local storage space per variant, will then be used instead of Aria, Opera’s native browser AI, until a user starts a new chat with the AI or switches Aria back on.

“Introducing Local LLMs in this way allows Opera to start exploring ways of building experiences and knowhow within the fast-emerging local AI space,” said Krystian Kolondra, EVP Browsers and Gaming at Opera.

Track record or innovation in the AI space

In early 2023, Opera presented Opera One, its AI-centric flagship browser based on modular design principles and a new browser architecture with a multithreaded compositor that allows for a smoother than ever processing of UX elements. Opera One features the Aria browser AI, which can be accessed via the browser sidebar of the browser command line. Aria is also available in the gamer-centric Opera GX, as well as in Opera browser on iOS and Android.

SOURCE: PRNewswire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img