Apple Develops Breakthrough Method for Running LLMs on iPhones (www.macrumors.com)
from throws_lemy@lemmy.nz to technology@lemmy.world on 21 Dec 2023 15:27
https://lemmy.nz/post/4678111

#technology

threaded - newest

AceSLS@ani.social on 21 Dec 2023 15:41 next collapse

This comment sums it up pretty nicely:

LOL innovative invention of swapping memory to storage…… maybe they can call it something cool like “cache”.

Apple being “innovative” my ass, lmao

cheese_greater@lemmy.world on 21 Dec 2023 15:51 next collapse

Cache u inside

jcg@halubilo.social on 23 Dec 2023 16:57 collapse

How bow dat?

Hegar@kbin.social on 21 Dec 2023 17:05 next collapse

The easiest way to tell that something's not really innovative is if the person describing it uses the word innovative.

HeartyBeast@kbin.social on 21 Dec 2023 19:02 collapse

Can you give an example of something that actually was innovative, that no-one called innovative?

Hegar@kbin.social on 21 Dec 2023 19:18 next collapse

The spoked wheel.

HeartyBeast@kbin.social on 21 Dec 2023 21:37 collapse

Interestingly, it looks as if nothing was really called innovative before 1960, with usage peaking in 2000, and its now in decline

https://www.etymonline.com/word/innovative

Hegar@kbin.social on 22 Dec 2023 07:29 collapse

Peaking in 2000 seems odd - I see or hear that word daily and that definitely wasn't the case back in 2000. Interesting!

GluWu@lemm.ee on 21 Dec 2023 21:14 collapse

Corndogs

lemmylommy@lemmy.world on 21 Dec 2023 20:05 next collapse

Well, if that commenter had more than just a vague idea of caching and/or swapping, they would know that the right algorithm can make or break performance.

That paper is not “we invented caching”, but “this is how we make some certain models work well despite constraints imposed by RAM and flash storage.”

It’s a worthy job for an engineer or researcher. Not quite as innovative as the invention of the wheel, but still enough to write a paper on (and read it, if you can manage to understand it).

4am@lemm.ee on 21 Dec 2023 21:30 collapse

35 upvotes in the technology community…man you guys really are just all knee-jerk reactionaries and it really knowledgeable tech at all. git gud

cheese_greater@lemmy.world on 21 Dec 2023 15:50 next collapse

Make Siri Great For Once

LazaroFilm@lemmy.world on 21 Dec 2023 16:25 next collapse

Huhum…?

Still working on that…

I’m sorry, try again later.

cheese_greater@lemmy.world on 21 Dec 2023 16:33 collapse

You’re triggering me lol

coolmojo@lemmy.world on 21 Dec 2023 16:56 collapse

Found the following websites about you’re triggering me lol.

cheese_greater@lemmy.world on 21 Dec 2023 17:11 collapse

Sigh! [unzips] Go ahead…

[deleted] on 21 Dec 2023 16:42 collapse

.

Lophostemon@aussie.zone on 21 Dec 2023 15:51 next collapse

Why the hell do we want to encourage people running MLMS on our phones?!! I don’t want to be part of some stupid pyramid scheme nonsense.

flooppoolf@lemmy.world on 21 Dec 2023 15:53 next collapse

Haha

LazaroFilm@lemmy.world on 21 Dec 2023 16:25 next collapse

dōTERRA Phone

Tronn4@lemmy.world on 21 Dec 2023 18:38 collapse

HerbaLife Galaxy S30

jsh@sh.itjust.works on 21 Dec 2023 16:46 next collapse

😭

Tja@programming.dev on 21 Dec 2023 19:47 next collapse

Here, you dropped this: /s

Lophostemon@aussie.zone on 21 Dec 2023 21:42 collapse

/s is for weaklings.

HEAR ME ROAR.

4am@lemm.ee on 21 Dec 2023 21:31 next collapse

Siri could suck an order of magnetize less and work offline, for starters

pineapplelover@lemm.ee on 22 Dec 2023 08:18 collapse

LLM≠MLM

Lophostemon@aussie.zone on 22 Dec 2023 09:48 collapse
guitarsarereal@sh.itjust.works on 21 Dec 2023 17:55 next collapse

Everyone likes to trash machine learning because the power requirements are high, but what they don’t realize is that we’re in the very first days of this technology (well, first couple decades of the technology being around, first few years of it being advanced enough to have anything to show off). Every technology that got bundled together into your phone was equally as useless when it was first invented. Honestly, compared to the development of most other technologies I’ve looked at, the pace of development in AI has been shocking.

Literally once a week, I see some news story about AI researchers delivering an order of magnitude speedup in some aspect of AI inference. The technique described here apparently allows for a 20x speedup on GPU’s.

cybersandwich@lemmy.world on 21 Dec 2023 18:03 collapse

Whispercpp works off the ML cores on the m series chips. It’s faster than my 1080ti that I have in a server doing the same things–by orders of magnitude. And it sips power.

Purpose built chips can be super powerful for their specific purposes.

fruitycoder@sh.itjust.works on 22 Dec 2023 08:23 next collapse

Tbh I’m more excited to see someone do use webnn, webgpu and petals together. Building smaller tighter models is good too.

wizzor@sopuli.xyz on 21 Dec 2023 16:40 collapse

I don’t understand the innovation, I already run LLMs and stable diffusion on a laptop from 2011.

I have no doubt it could be run on my Android phone.