cmnybo@discuss.tchncs.de
on 04 Apr 2024 03:06
nextcollapse
I haven’t used Opera since they switched from their own engine to chrome. They are now owned by a Chinese company, so it probably has at least as much tracking built into it as Google Chrome now.
rottingleaf@lemmy.zip
on 04 Apr 2024 07:13
nextcollapse
I used Fifth a bit, which is something aesthetically similar to old Opera made with fltk and a webkit port to fltk. But it’s abandoned now.
It’s so sad really, when I was a Windows user, it was Opera, when I moved to Linux, it was again Opera, then I also started using Conkeror (based on XULRunner).
Then Opera died. Then XULRunner died. No usable web browser anymore.
SharkAttak@kbin.social
on 04 Apr 2024 18:49
collapse
Don't you like Firefox? It's on both win and Linux.
rottingleaf@lemmy.zip
on 04 Apr 2024 20:25
collapse
I said “usable” ; it was usable when XULRunner was a thing (and you could use Firefox instead of just XULRunner).
coolmojo@lemmy.world
on 04 Apr 2024 13:05
collapse
Have a look at Otter browser It aims to replicate the old interface. It is using QtWebEngine as Presto was closed source. It is in development since 10 years now. And it is open source.
squid_slime@lemm.ee
on 04 Apr 2024 15:14
nextcollapse
coolmojo@lemmy.world
on 05 Apr 2024 18:34
collapse
Qt WebEngine uses code from the Chromium project. However, it is not containing all of Chrome/Chromium:
Binary files are stripped out
Auxiliary services that talk to Google platforms are stripped out,
Source
While that’s one of the reasons I don’t want to use chromium, it’s not actually the main reason, if so I’d just use Ungoogled Chromium. I just want more web engines, and I dont want google to monopolise the internet.
coolmojo@lemmy.world
on 06 Apr 2024 11:13
collapse
It is super hard to create a new web engine, especially when one company is influencing the web standards and most web developers are only testing against that because of market share. This is why we ended up with four active web engines. In alphabetical order: Blink, Gecko, Goanna, WebKit. Obviously some are related: WebKit started out as the fork of KDE’s KHTML and Blink is the fork of WebKit. Goanna is the fork of the Unified XUL Platform that was forked from Mozilla’s Gecko.
reddig33@lemmy.world
on 04 Apr 2024 03:25
nextcollapse
Why the hell do I need this in a web browser? Why isn’t it a stand alone app?
FaceDeer@fedia.io
on 04 Apr 2024 04:26
nextcollapse
There are plenty of stand-alone LLM apps.
GlitterInfection@lemmy.world
on 04 Apr 2024 04:58
nextcollapse
If you think of LLMs as a thing to replace search bars then this kind of makes sense.
reddig33@lemmy.world
on 04 Apr 2024 05:06
nextcollapse
Just more unnecessary browser bloat.
GlitterInfection@lemmy.world
on 04 Apr 2024 05:06
collapse
The more search bars the faster your internet becomes!
GlitterInfection@lemmy.world
on 04 Apr 2024 05:23
collapse
This is true. I asked my LLM.
noodlejetski@lemm.ee
on 04 Apr 2024 08:35
collapse
If you think of LLMs as a thing to replace search bars
I don’t.
GlitterInfection@lemmy.world
on 04 Apr 2024 21:15
collapse
I haven’t tried LLMs myself, but even completely made up garbage would be better than today’s search engine results.
You either get advertisements for things that have nothing to do with what you’re trying to find or you get privacy preserving links to sites that have nothing to do with what you’re trying to find.
chronicledmonocle@lemmy.world
on 04 Apr 2024 05:07
collapse
Same reason people get their WiFi from their ISP Modem+Router combo, even though it’s stupid to do so: People often confuse initial convenience for good.
gunpachi@lemmings.world
on 04 Apr 2024 03:43
nextcollapse
Thats a cool feature for sure but I don’t trust opera.
Bandicoot_Academic@lemmy.one
on 04 Apr 2024 05:16
nextcollapse
Intresting. But I’m curious about the performance.
A bigger LLM (mixtral) already struggles to run on my mid-range gaming PC. Trying to run an LLM that isn’t terrible on a standard laptop wouldn’t be a good experience.
I have no idea how this is set up to work technically, but most of the heavy lifting is gonna be on the GPU. I’m not sure that it matters much whether the browser is what’s pushing data to the GPU or some other package.
Bandicoot_Academic@lemmy.one
on 04 Apr 2024 06:39
nextcollapse
Most people probably don’t have a dedicated GPU and an iGPU is probably not powerfull enough to run an LLM at decent speed. Also a decent model requires like 20GB of RAM which most people don’t have.
douglasg14b@lemmy.world
on 04 Apr 2024 07:56
collapse
It doesn’t just require 20GB of RAM, it requires that in VRAM. Which is a much higher barrier to entry.
Hamartiogonic@sopuli.xyz
on 04 Apr 2024 10:16
collapse
But what if you have an AMD APU. Doesn’t that use your normal RAM as VRAM?
Not exactly. Most integrated chips have a small pool of dedicated VRAM, and then a bit more that they share with the system memory, though it’s generally only a portion, not all of it. It’s only Apple’s unified memory, and maybe other mobile chips that has them both share memory pool entirely, for better or worse, as far as I’m aware.
But it is worth noting that if you don’t have enough VRAM and have to put it into RAM, the minimum expectation is that you have twice the amount of RAM space. So if you have a GPU with 4GB of VRAM, and need to offload the extra to the system, you don’t need 16 GB, you need 32 GB.
T156@lemmy.world
on 04 Apr 2024 07:00
nextcollapse
Unlikely, at least on non-nvidia chips, and even on AMD, it’s only the latest four chips that support it. Anything older isn’t going to cut it.
You also need a fairly big amount of VRAM for models like that. (4 GB is the minimum for the common kinds, which is more than typical integrated systems, or 8 GB of system memory). You can get by with system RAM, but the performance will be quite bad, since you’re either relying on the CPU, or you’ll be adding the latency from data moving between them.
The thought that internet becomes shitty enough that you need a GPU to browse it is really frightening me. If we really reach that point that may be to run an AI which filters out AI generated spam which would really depress me 😭
The thought that internet becomes shitty enough that you need a GPU to browse it is really frightening me.
I mean, there was a point where an FPU was a separate chip and wasn’t the norm; now it’s built into the CPU.
I think that it’s probably safe to say that, in the future, there will be broader use of parallel processing, as we’ve fundamental limits on what we know we can do there with existing laws of physics with serial processing. That could wind up being part of the CPU. It could live on a separate piece of hardware – which may not necessarily be a “GPU” – parallel processing hardware entered the PC because the most-immediate need was to do 3d graphics rendering, but as you can see from the LLMs that people are running on GPUs today, that’s not the only application. The parallel compute accelerator cards that Nvidia is selling today for an arm and a leg on servers aren’t aimed at doing 3d graphics.
It may not be 3d graphics rendering or running LLMs that becomes the primary application. But I’d be reasonably comfortable saying that down the line, relative to today, there will be more parallel-processing hardware in computers than is present today.
For a lot of newer processors, it already is. Intel, Apple, Samsung, and Qualcomm like to brag about their processors having some sort of neural contraption meant to assist with AI processing.
If it stays around, it might be good enough that you don’t need a GPU to do it, since the CPU has an onboard chip that can handle that work instead, since tensor processors like that are a bit more efficient than GPUs, but are also more specialised.
essteeyou@lemmy.world
on 04 Apr 2024 05:25
nextcollapse
Can’t they just stick to normal browser things like gaming integrations?
folak@lemmy.world
on 04 Apr 2024 06:32
nextcollapse
threaded - newest
I haven’t used Opera since they switched from their own engine to chrome. They are now owned by a Chinese company, so it probably has at least as much tracking built into it as Google Chrome now.
I used Fifth a bit, which is something aesthetically similar to old Opera made with fltk and a webkit port to fltk. But it’s abandoned now.
It’s so sad really, when I was a Windows user, it was Opera, when I moved to Linux, it was again Opera, then I also started using Conkeror (based on XULRunner).
Then Opera died. Then XULRunner died. No usable web browser anymore.
Don't you like Firefox? It's on both win and Linux.
I said “usable” ; it was usable when XULRunner was a thing (and you could use Firefox instead of just XULRunner).
I miss old opera before the buyout
That’s essentially Vivaldi now.
Apart from it being chromium based 😕
Have a look at Otter browser It aims to replicate the old interface. It is using QtWebEngine as Presto was closed source. It is in development since 10 years now. And it is open source.
Thanks, didn’t know about this
QtWebEngine is Chromium :(
It’s Chromium all the way down.
Qt WebEngine uses code from the Chromium project. However, it is not containing all of Chrome/Chromium: Binary files are stripped out Auxiliary services that talk to Google platforms are stripped out, Source
While that’s one of the reasons I don’t want to use chromium, it’s not actually the main reason, if so I’d just use Ungoogled Chromium. I just want more web engines, and I dont want google to monopolise the internet.
It is super hard to create a new web engine, especially when one company is influencing the web standards and most web developers are only testing against that because of market share. This is why we ended up with four active web engines. In alphabetical order: Blink, Gecko, Goanna, WebKit. Obviously some are related: WebKit started out as the fork of KDE’s KHTML and Blink is the fork of WebKit. Goanna is the fork of the Unified XUL Platform that was forked from Mozilla’s Gecko.
Why the hell do I need this in a web browser? Why isn’t it a stand alone app?
There are plenty of stand-alone LLM apps.
If you think of LLMs as a thing to replace search bars then this kind of makes sense.
Just more unnecessary browser bloat.
Like search bars.
The more search bars the faster your internet becomes!
This is true. I asked my LLM.
I don’t.
I haven’t tried LLMs myself, but even completely made up garbage would be better than today’s search engine results.
You either get advertisements for things that have nothing to do with what you’re trying to find or you get privacy preserving links to sites that have nothing to do with what you’re trying to find.
Same reason people get their WiFi from their ISP Modem+Router combo, even though it’s stupid to do so: People often confuse initial convenience for good.
Thats a cool feature for sure but I don’t trust opera.
Intresting. But I’m curious about the performance.
A bigger LLM (mixtral) already struggles to run on my mid-range gaming PC. Trying to run an LLM that isn’t terrible on a standard laptop wouldn’t be a good experience.
I have no idea how this is set up to work technically, but most of the heavy lifting is gonna be on the GPU. I’m not sure that it matters much whether the browser is what’s pushing data to the GPU or some other package.
Most people probably don’t have a dedicated GPU and an iGPU is probably not powerfull enough to run an LLM at decent speed. Also a decent model requires like 20GB of RAM which most people don’t have.
It doesn’t just require 20GB of RAM, it requires that in VRAM. Which is a much higher barrier to entry.
But what if you have an AMD APU. Doesn’t that use your normal RAM as VRAM?
Not exactly. Most integrated chips have a small pool of dedicated VRAM, and then a bit more that they share with the system memory, though it’s generally only a portion, not all of it. It’s only Apple’s unified memory, and maybe other mobile chips that has them both share memory pool entirely, for better or worse, as far as I’m aware.
But it is worth noting that if you don’t have enough VRAM and have to put it into RAM, the minimum expectation is that you have twice the amount of RAM space. So if you have a GPU with 4GB of VRAM, and need to offload the extra to the system, you don’t need 16 GB, you need 32 GB.
Unlikely, at least on non-nvidia chips, and even on AMD, it’s only the latest four chips that support it. Anything older isn’t going to cut it.
You also need a fairly big amount of VRAM for models like that. (4 GB is the minimum for the common kinds, which is more than typical integrated systems, or 8 GB of system memory). You can get by with system RAM, but the performance will be quite bad, since you’re either relying on the CPU, or you’ll be adding the latency from data moving between them.
The thought that internet becomes shitty enough that you need a GPU to browse it is really frightening me. If we really reach that point that may be to run an AI which filters out AI generated spam which would really depress me 😭
I mean, there was a point where an FPU was a separate chip and wasn’t the norm; now it’s built into the CPU.
I think that it’s probably safe to say that, in the future, there will be broader use of parallel processing, as we’ve fundamental limits on what we know we can do there with existing laws of physics with serial processing. That could wind up being part of the CPU. It could live on a separate piece of hardware – which may not necessarily be a “GPU” – parallel processing hardware entered the PC because the most-immediate need was to do 3d graphics rendering, but as you can see from the LLMs that people are running on GPUs today, that’s not the only application. The parallel compute accelerator cards that Nvidia is selling today for an arm and a leg on servers aren’t aimed at doing 3d graphics.
It may not be 3d graphics rendering or running LLMs that becomes the primary application. But I’d be reasonably comfortable saying that down the line, relative to today, there will be more parallel-processing hardware in computers than is present today.
For a lot of newer processors, it already is. Intel, Apple, Samsung, and Qualcomm like to brag about their processors having some sort of neural contraption meant to assist with AI processing.
If it stays around, it might be good enough that you don’t need a GPU to do it, since the CPU has an onboard chip that can handle that work instead, since tensor processors like that are a bit more efficient than GPUs, but are also more specialised.
Can’t they just stick to normal browser things like gaming integrations?
rentry.co/operagx
What’s LLM?
en.wikipedia.org/wiki/Large_language_model
A lot of the “AI” stuff that’s been in the news recently, chatbots and image generation and such, are based on LLMs.