218

They support Claude, ChatGPT, Gemini, HuggingChat, and Mistral.

you are viewing a single comment's thread
view the rest of the comments
[-] TheMachineStops@discuss.tchncs.de 4 points 3 weeks ago* (last edited 3 weeks ago)

It gives you many options on what to use, you can use Llama which is offline. Needs to be enabled though about:config > browser.ml.chat.hideLocalhost.

[-] Swedneck@discuss.tchncs.de 5 points 3 weeks ago

and thus is unavailable to anyone who isn't a power user, as they will never see a comment like this and about:config would fill them with dread

[-] TheMachineStops@discuss.tchncs.de 4 points 3 weeks ago* (last edited 3 weeks ago)

Lol, that is certainly true and you would need to also set it up manually which even power users might not be able to do. Thankfully there is an easy to follow guide here: https://ai-guide.future.mozilla.org/content/running-llms-locally/.

[-] LWD@lemm.ee 1 points 3 weeks ago

There's a huge difference between something that is presented in an easily accessible settings menu, and something that requires you to go to an esoteric page, click through a scary warning message, and then search for esoteric settings... Before even installing a server.

Nothing was compelling Mozilla to rush this through. In addition, nobody was asking Mozilla for remote access to AI, AFAIK. Before Mozilla pushed for it, people were praising them for resisting the temptation to follow the flock. They could have waited and provided better defaults.

Or just wedged it into an extension, something they're currently doing anyway.

this post was submitted on 27 Nov 2024
218 points (100.0% liked)

Firefox

18045 readers
6 users here now

A place to discuss the news and latest developments on the open-source browser Firefox

founded 5 years ago
MODERATORS