The problem with Firefox doing AI is theyre one foot out always. The features they add are always undercooked compared to the rest of the market. This looks really shit and useless in its current state like a worse version of perplexity browser.
All AI is undercooked: Errors are baked into LLMs and there is no viable solution to prevent the mistakes and outright bullshit they produce other than to assume it fucked up and pay an actual expert manually check literally everything it does.
Errors are baked in but I don’t agree with the “no viable solution” part. One research team actually was able to identify the “neurons” responsible for hallucinations and adjust the contribution to negligible amounts.
https://www.youtube.com/watch?v=1ONwQzauqkc
(Linking a youtuber instead of the actual study because he summarizes it pretty well and the research itself is not geared for laypersons.)
If this was implemented industry wide would it completely solve the problem? I don’t know, but I do know it would be a massive improvement.
I never took what you wrote to mean that, but I am deeply skeptical that they can successfully elminiate hallucinations to the point that “AI” can be trusted to given correct results.
Why bring up power and environmental cost? What did that have to do with anything?
Also if you’ll re-read what I wrote I used careful language to indicate I didn’t think this method would completely eliminate errors. Nevermind bridge the gap to “trusted.” (🤮 I will never trust AI.)
(Yeah I know the YouTuber used a sensational title; in their defense they kind of have to in order to get clicks. imho blame the algorithm and people’s reinforcement of that algorithm.)
Why wouldn’t I? It’s pretty fucking important! Why would you take exception to that? I also think it’s weird you assumed what conclusion I was jumping to.
The problem with Firefox doing AI is theyre one foot out always. The features they add are always undercooked compared to the rest of the market. This looks really shit and useless in its current state like a worse version of perplexity browser.
All AI is undercooked: Errors are baked into LLMs and there is no viable solution to prevent the mistakes and outright bullshit they produce other than to assume it fucked up and pay an actual expert manually check literally everything it does.
Errors are baked in but I don’t agree with the “no viable solution” part. One research team actually was able to identify the “neurons” responsible for hallucinations and adjust the contribution to negligible amounts.
https://www.youtube.com/watch?v=1ONwQzauqkc (Linking a youtuber instead of the actual study because he summarizes it pretty well and the research itself is not geared for laypersons.)
If this was implemented industry wide would it completely solve the problem? I don’t know, but I do know it would be a massive improvement.
I remain deeply skeptical.
Either way, it uses a ridiculous amount of power and comes at great environmental cost.
Fuck me, you and people in general jump to conclusions so easily. My post was meant to educate, to shore up knowledge. To help out.
In no way was I saying “AI is good and the tech bros are right about it.” 🤦♂️
I never took what you wrote to mean that, but I am deeply skeptical that they can successfully elminiate hallucinations to the point that “AI” can be trusted to given correct results.
Why bring up power and environmental cost? What did that have to do with anything?
Also if you’ll re-read what I wrote I used careful language to indicate I didn’t think this method would completely eliminate errors. Nevermind bridge the gap to “trusted.” (🤮 I will never trust AI.)
(Yeah I know the YouTuber used a sensational title; in their defense they kind of have to in order to get clicks. imho blame the algorithm and people’s reinforcement of that algorithm.)
Why wouldn’t I? It’s pretty fucking important! Why would you take exception to that? I also think it’s weird you assumed what conclusion I was jumping to.