They support Claude, ChatGPT, Gemini, HuggingChat, and Mistral.
Didn’t want it in Opera, don’t want it in Firefox. I mean they can keep trying and I’ll just keep on ignoring this shit :/
This happened ages ago, didn’t it? Am I missing something new?
If they do it in a privacy-preseeving way, this could help them get back market share which will generally benefit an open internet.
oh good. hurray.
I will say, the Le Chat provider is pretty decent. You really can use it more natural language. “Rewrite it with a better rhyme scheme” “remove the last line” and it just got it.
Why no local option though? Why no anonmysing option? Why no right click use here option for different text fields? (Like help write this as better auto complete or summarize this block of text, etc) Why no RAG using the web, why only generated text?
I mean, if you’re going to do it, where’s the Ollama love?
I was disappointed there was no local option…
I don’t get it, ollama is a provider no?
I think the point is it’s open source
and so is firefox, so why use another model provider
I don’t understand the hate. It’s just a sidebar for the supported LLMs. Maybe I’m misunderstanding?
Yes, I would prefer Mozilla focus on the browser, but to me, this seems like it was done in an afternoon.
Now add support for GPT4All and everyone is happy again.
Thanks for nothing, Mozilla.
They should raise the ceo’s pay some more to celebrate.
And fire a few employees just cause.
Thing is, for your average user with no GPU and whp never thinks about RAM, running a local LLM is intimidating. But it shouldn’t be. Any system with an integrated GPU, and the more RAM the better, can run simple models locally.
The not so dirty secret is that ChatGPT 3 vs 4 isn’t that big a difference, and neither are leaps and bounds ahead of the publically available models for about 99% of tasks. For that 1% people will ooh and aah over it, but 99% of use cases are only seeing marginal gains on 4o.
And the simplified models that run “only” 95% as well? They can use 90% fewer resources give pretty much identical answers outside of hyperspecific use cases.
Running a a “smol” model as some are called, gets you all the bang for none of the buck, and your data stays on your system and never leaves.
I’ve been yelling from the rooftops to some stupid corporate types that once the model is trained, it’s trained. Unless you are training models yourself, there is no need for the massive AI clusters, just for the model. Run it local on your hardware at a fraction of the cost.
There’s the tragedy with this new feature: they fast-tracked this past more popular requests, sticking it into Release Firefox.
But they only rushed the part that connects to third parties. There was also a “localhost” option which was originally alongside the Big Five corporate offerings, but Mozilla ultimately decided to bury that one inside of the
about:config
settings.I’m guessing that the reason (and a good one at that) is that simply having an option to connect to a local chatbot leads to just confused users because they also need the actual chatbot running on their system. If you can set up that, then you can certainly toggle a simple switch in about:config to show the option.
Can you point me to some resources to running smol llm?
My use case prob just to help “typing” miscellaneous idea I have or check for my grammatical error, in english.
Thanks, in advance.
Here you go: Review of SmolVLM https://www.marktechpost.com/2024/11/26/hugging-face-releases-smolvlm-a-2b-parameter-vision-language-model-for-on-device-inference/
Model itself: https://huggingface.co/spaces/HuggingFaceTB/SmolVLM
And you can use Ollama to run it locally, and Open WebUI to access it in browser.
Last time I tried using a local llm (about a year ago) it generated only a couple words per second and the answers were barely relevant. Also I don’t see how a local llm can fulfill the glorified search engine role that people use llms for.
Try again. Simplified models take the large ones and pare them down in terms of memory requirements, and can be run off the CPU even. The “smol” model I mentioned is real, and hyperfast.
Llama 3.2 is pretty solid as well.
They’re fast and high quality now. ChatGPT is the best, but local llms are great, even with 10gb of vram.
For a second I thought it said “experimental failure”. Would be more accurate, I think.
why a fucking chatbot? translate a page better for me you fucking losers, all the translation options suck for privacy outside of specifically trained local AIs. this is the BEST use case for a small local LLM yet mozilla with all its brains and resources couldnt rub two neurons together for this.
or they could do character prediction on your typing to make typing faster. just some legit examples, why waste resources to build a chat ai into my browser when i can just open a website???
Perhaps Mozilla’s biggest “failure” is just communication…
Firefox actually has this now.
Could this replace Perplexity for (assisted free) online search?
I wish I had telemetry on such features.
I really doubt a significant number of people use AI chatbots often enough that having it in a dedicated sidebar is worth it.
I’ve never had the urge to use a chat bot personally, but I’m pretty sure I’m in the minority. Lots of people use these things all the time for so much stuff we probably wouldn’t even consider.
I’ve worked with a few people that all but rely on these things to produce any creative work they have to do.
Maybe we run in different circles but I think a lot of people don’t even talk about how they’re using it.
I wish I had telemetry
I’m sure they do as Mozilla is an ad company
This is apparently either not widely known or some people just like to shoot the messenger.
- jwz, Jun.: Mozilla is an advertising company now
- jwz, Oct.: Mozilla’s CEO doubles down on them being an advertising company now
- Mozilla support: Share data with Mozilla to help improve Firefox
- Firefox documentation: Telemetry