Skip to content

Opera unveils support for local LLMs within the browser

  • by
  • 3 min read

Opera has introduced a new feature allowing users to download 150 local Large Language Model (LLM) variants from around 50 model families directly within its browser through the AI Feature Drops program.

The newly integrated local LLM variants include popular models such as Meta’s Llama, Vicuna, Gemma from Google, and Mixtral from Mistral AI. These variants are now accessible in Opera One’s developer stream, offering users enhanced control and confidentiality over their AI interactions.

This advancement represents a unique integration of AI within web browsers. By incorporating local LLMs, Opera empowers users to utilise AI capabilities while safeguarding the data.

This is an image of dropdownwithmodels opera llm ss1
Source: Opera

Users can easily access these local LLMs within Opera One’s developer stream and select their preferred model for prompt processing, considering individual preferences and performance needs. However, it is important to note that local LLMs may require varying storage capacities (2-10 GB) and processing speeds based on users’s hardware.

“Using local large language models means users’ data is kept locally, on their device, allowing them to use AI without the need to send information to a server. We are testing this new set of local LLMs in Opera One’s developer stream as part of our AI Feature Drops Program, which allows you to test early, often experimental versions of our AI feature set,” noted Julia Szyndzielorz, Senior Public Relations Manager at Opera.

This is an image of operallm ss1
Source: Opera

If the users’s processor speed is slow, they may feel some delay in getting the answers compared to a server-based LLM.

To test the feature, users must update Opera to the latest version, open Aria Chat, and click on Choose local mode. Then, they must search for and download the version they like. After finishing the download, they must click the menu button, select Choose local mode again, and then select the model they have downloaded to start the chat.

This initiative also paves the way for potential future use cases, such as browsers leveraging AI solutions based on users’ historical input while preserving data locally, signalling a personalised browsing experience.

In the News: TSMC and UMC halt production following Taiwan earthquake 

Kumar Hemant

Kumar Hemant

Deputy Editor at Candid.Technology. Hemant writes at the intersection of tech and culture and has a keen interest in science, social issues and international relations. You can contact him here: kumarhemant@pm.me

>