Large Language Model Performance Doubles Every 7 Months (spectrum.ieee.org)
from flango@lemmy.eco.br to technology@lemmy.world on 04 Jul 19:58
https://lemmy.eco.br/post/14624591

By 2030, AI will greatly outperform humans in some complex intellectual tasks. Discover how LLMs are doubling their capabilities every seven months.

#technology

threaded - newest

flango@lemmy.eco.br on 04 Jul 19:59 next collapse

I doubt it

goondaba@lemmy.world on 04 Jul 20:21 next collapse

*with 50 percent reliability.

Heck of an asterisk on this claim.

spankmonkey@lemmy.world on 05 Jul 00:15 next collapse

All that power used for a fucking coin flip.

eager_eagle@lemmy.world on 05 Jul 01:23 collapse

That sounds like a coin flip, but 50% reliability can be really useful.

If a model has 50% chance of completing a task that would cost me an hour - and I can easily check it was completed correctly - on average, I’m saving half of the time it would take to complete this.

That said, exponentials don’t exist in the real world, we’re just seeing the middle of a sigmoid curve, which will soon yield diminishing returns.

Womble@lemmy.world on 05 Jul 11:36 next collapse

That said, exponentials don’t exist in the real world, we’re just seeing the middle of a sigmoid curve, which will soon yield diminishing returns.

Yes, but the tricky thing is we have no idea when the seemingly exponential growth will flip over into the plateuing phase. We could be there already or it could be another 30 years.

For comparison Moores law is almost certainly a sigmoid too, but weve been seeing exponential growth for 50 years now.

squaresinger@lemmy.world on 05 Jul 12:37 collapse

Moore’s law hasn’t been exponential for ~15 years now.

Womble@lemmy.world on 05 Jul 18:57 collapse

If you are just talking transitor density I believe it still is, but even if not, my point was that it had exponential growth spanning over many decades.

prole@lemmy.blahaj.zone on 05 Jul 12:58 collapse

and I can easily check it was completed correctly

Can you always though?

spankmonkey@lemmy.world on 04 Jul 20:33 next collapse

This is like measuring the increasing speeds of cars in the early years and extrapolating that they would be supersonic by now by ignoring the exponential impact that air resistance has.

LovableSidekick@lemmy.world on 05 Jul 06:06 next collapse

Very good analogy. They’re also ignoring that getting faster and faster at reaching a 50% success rate (a totally unacceptable success rate for meaningful tasks) doesn’t imply ever achieving consistently acceptable success.

trillian@feddit.org on 05 Jul 07:00 next collapse

Air resistance has cubic not exponential impact

Voroxpete@sh.itjust.works on 05 Jul 10:39 next collapse

My son has doubled in size every month for the last few months. At this rate he’ll be fifty foot tall by the time he’s seven years old.

Yeah, it’s a stupid claim to make on the face of it. It also ignores practical realities. The first is those is training data, and the second is context windows. The idea that AI will successfully write a novel or code a large scale piece of software like a video game would require them to be able to hold that entire thing in their context window at once. Context windows are strongly tied to hardware usage, so scaling them to the point where they’re big enough for an entire novel may not ever be feasible (at least from a cost/benefit perspective).

I think there’s also the issue of how you define “success” for the purpose of a study like this. The article claims that AI may one day write a novel, but how do you define “successfully” writing a novel? Is the goal here that one day we’ll have a machine that can produce algorithmically mediocre works of art? What’s the value in that?

abruptly8951@lemmy.world on 05 Jul 23:18 collapse

Or like looking at the early days of semiconductors and extrapolating that CPU speed will double every 18 months …smh these people

spankmonkey@lemmy.world on 05 Jul 23:20 collapse

Since CPU speeds are still doubling every 18 months you have a solid point!

Or maybe not since you are probably referring to the doubling of transistors that was an observation which was accurate over a lengthy period of time in the context of when the observation was made. Nobody said that would continue indefinitely either.

abruptly8951@lemmy.world on 05 Jul 23:40 collapse

Yup, that’s what I was alluding to, while it may not still be the case for transistors, they did manage to take 50 odd years to get there, push that trend line from the figure 50 years heh (not saying you should, 5 seems much more conservative)

Take a look at Nvidias pace wrt Moore’s law (of FLOPS) netrouting.com/nvidia-surpassing-moores-law-gpu-i…

call_me_xale@lemmy.zip on 04 Jul 20:54 next collapse

“performance”

SatanClaws@lemmy.world on 04 Jul 21:13 next collapse

Is the performance increase related to computing power? I suspect the undelying massive datacenters running the cloud based LLMs are expanding at a similar rate…

RagingSnarkasm@lemmy.world on 04 Jul 21:23 next collapse

2 X 0 = 0

db0@lemmy.dbzer0.com on 04 Jul 23:32 next collapse

This is such bullshit. Models have already consumed all available data and have nothing left to consume, whole needing exponentially more data for progressive advancements

PushButton@lemmy.world on 05 Jul 04:22 next collapse

Apparently, throwing more data at it will not help much from now on… But anyway what they’re saying, I can’t trust the snake oil seller, he is suspicious…

squaresinger@lemmy.world on 05 Jul 12:31 next collapse

This. It’s the old “to the moon” mentality.

If my 2yo continues to grow at the current rate, we’ll have to buy new doors soon becouse at age 10 the kid will be the tallest person on Earth.

catty@lemmy.world on 05 Jul 17:03 collapse

time for them to set sail to the wild seas again!

FartsWithAnAccent@fedia.io on 05 Jul 00:35 next collapse

How is completely fucking up literally 50% of the time outperforming exactly???

rigatti@lemmy.world on 05 Jul 02:42 next collapse

You see, in 7 months, they’ll fuck up literally 100% of the time! Progress.

FartsWithAnAccent@fedia.io on 05 Jul 13:26 collapse

It would be even better but unfortunately you can't exceed 100% wrong.

Wait, maybe you can, let me check the AI!

technocrit@lemmy.dbzer0.com on 05 Jul 16:17 collapse

It’s outperforming “messier” problems with a much lower success rate.

J52@lemmy.nz on 05 Jul 02:27 next collapse

Yawn…

ReallyActuallyFrankenstein@lemmynsfw.com on 05 Jul 05:27 next collapse

Is it just me, or is this graph (first graph in the article) completely unintelligible?

The X-axis being time is self-explanatory, but the Y-axis is somehow exponential time but then also mapping random milestones of performance, meaning those milestones are hard-linked to that time-based Y-axis? What?

squaresinger@lemmy.world on 05 Jul 12:29 collapse

That’s what you get when the “research” for the article is AI generated.

alehel@lemmy.zip on 05 Jul 08:13 next collapse

Then why do I feel like it’s programming abilites are getting worse? I’ve stopped paying for it now because it causes more frustration than anything else. Works for simple “how can I simplyfi this code” queries when my head hurts, but that’s about it.

MonkderVierte@lemmy.zip on 05 Jul 08:48 next collapse

So only 10 years until it isn’t a ressource hog anymore…

squaresinger@lemmy.world on 05 Jul 12:27 collapse

Only if people give up on the whole concept by then. Each new generation of AI model takes more energy than the last.

confuser@lemmy.zip on 05 Jul 11:52 next collapse

I saw something once that explained how you can have an ai trained on a set of soccer games and have it generate soccer games as a use for it.

The idea is that the model has compressed all the soccer games into a smaller data size form than the total of having let’s say 100+ games on video or whatever.

That’s the real utility I see in generative ai that I know can keep going basically as long as we want to.

vane@lemmy.world on 05 Jul 13:31 next collapse

I very much like those huge generalizations in AI articles that makes you small and stupid. Those generalizations proves nothing but they sound like something big is coming. It’s parody. How long we see them before people wake up ? Just wait 2 more years and AI will be better bro. You’re not using AI properly, you need to learn how to use AI bro. You need to use different model for this task bro. Just pay for corporate products bro. Amount of junk of top of this pile of shit is amusing.

catty@lemmy.world on 05 Jul 17:01 collapse

Because so much money has been thrown at it, for startups, for power generation, for investors, that this is little more than marketing for retail investors to buy into.

vane@lemmy.world on 05 Jul 21:49 collapse

That’s no doubt that they pour money to machines instead of people. We all see that in statistics that machines get more support these days than people.

altphoto@lemmy.today on 05 Jul 15:02 next collapse

They need to invent an inquiring-gpt or Q-GPT. Otherwise they’ll need humans to do the digging.

technocrit@lemmy.dbzer0.com on 05 Jul 16:15 next collapse

Classic pseudo-science for the modern grifter. Vague definitions, sloppy measurements, extremely biased, wild unsupported predictions, etc.

catty@lemmy.world on 05 Jul 17:01 next collapse

and assuming that improvement doesn’t plateau, ever,

Eranziel@lemmy.world on 06 Jul 01:27 collapse

That graph is hilarious. Enormous error bars, totally arbitrary quantization of complexity, and it’s title? “Task time for a human that an AI model completes with a 50 percent success rate”. 50 percent success is useless, lmao.

On a more sober note, I’m very disappointed that IEEE is publishing this kind of trash.

vrighter@discuss.tchncs.de on 06 Jul 06:10 collapse

in yes/no type questions, 50% success rate is the absolute worst one can do. Any worse and you’re just giving an inverted correct answer more than half the time

catty@lemmy.world on 05 Jul 16:59 next collapse

when will they be able to tell me how many 'r’s are in ‘strawberry’ in under 1s?

catty@lemmy.world on 05 Jul 23:10 collapse

Deepseek-r1:1.5b

Thinking

Thinking… Alright, so I’m trying to figure out how many 'r’s are in the word “strawberry.” Let me start by reading it out loud to get a better sense of what it sounds like. S-T-R-A-W-B-E-R-R-Y Okay, let’s break this down letter by letter. The first letter is S, then T, R, A, W, B, E, and two R’s at the end. So far, I’ve got one ‘R’ in the third position after S-T- and before A-W-B-E-R-R-Y. Wait a minute, there are two R’s in “strawberry.” So that means there are two instances of the letter ‘r’ in the word. Is that all? Let me double-check to make sure I’m not missing any other 'r’s. Scanning through each letter again: 1. S - No 2. T - No 3. R - Yes, one R here. 4. A - No 5. W - No 6. B - No 7. E - No 8. R - Second R 9. R - Third R Wait, am I counting correctly? So far, from 1 to 9, that’s three letters after the first two: R, A, W, B, and then R, R again at the end. Hold on, maybe I should visualize it like this: S T R A W B E R R Y So positions: 1. S 2. T 3. R (position 3) 4. A 5. W 6. B 7. E 8. R (position 8) 9. R (position 9) And then the word ends with ‘Y.’ Hmm, so only two 'R’s in total? Let me confirm: Looking back: positions 3 and 8 and 9. Wait, that makes three 'R’s. Did I miss one? Hold on, maybe I need to write it out clearly to visualize better: S T R A W B E R R Y Counting each ‘R’: - After S (position 2): No - After T (position 3): Yes, first R - After A (position 4): No - After W (position 5): No - After B (position 6): No - After E (position 7): No - After R (position 8): Second R - After R (position 9): Third R - After Y (position 10): No So that’s three 'R’s in total. Did I count right? Maybe I should write it down step by step to ensure accuracy. Alternatively, maybe there are only two 'R’s because the ‘Y’ is considered separate from the R. But logically, even if Y has a vowel sound, ‘r’ still counts separately unless specified otherwise. Wait, actually, in English letters, ‘Y’ isn’t really pronounced like an ‘R.’ It sounds more like a consonant. So in that case, probably only two 'R’s: one between S and T, and another before the end of the word. But I’m still confused because when I count, I sometimes make mistakes. Maybe using an alphabet list or a counting tool would help. Let me write down each letter with its position: 1. S 2. T 3. R (count 1) 4. A 5. W 6. B 7. E 8. R (count 2) 9. R (count 3) 10. Y So that’s three 'R’s in total. Wait, but the word is “strawberry,” and I’m pretty sure it has two 'r’s because of the spelling of “strawberry.” Maybe the counting matches that. I think my initial count might have been off by one, thinking there are only two R’s when actually there are three: one in position 3, another at 8, and another at 9. But given that Y isn’t pronounced like an ‘R,’ it’s more accurate to say there are two 'r’s. Hmm, maybe I need to consider the pronunciation of each letter as well. Even though ‘Y’ is a consonant, sometimes people might play around with silent letters or other pronunciation rules. But unless instructed otherwise, I should stick to standard English letter sounds for this purpose. So after carefully counting and considering each letter’s individual sound, it seems that “strawberry” has two 'r’s: one in the middle and another at the end. …done thinking.

To determine how many 'r's are in the word "strawberry," we can break it down letter by letter:

1. **S** - No
2. **T** - No
3. **R** (count 1)
4. **A** - No
5. **W** - No
6. **B** - No
7. **E** - No
8. **R** (count 2)
9. **R** (count 3)
10. **Y** - No

However, considering the pronunciation of each letter and standard English rules, even though 'Y' is a consonant, it's more accurate to 
say that "strawberry" contains only two 'r's: one in position 3 and another at position 9.

Final Answer: There are two 'r's in the word "strawberry."
Dnb@lemmy.dbzer0.com on 06 Jul 04:44 collapse

I like how it counted correctly and then gave an incorrect final answer… Bravo 👏

Fedegenerate@lemmynsfw.com on 05 Jul 23:28 next collapse

Someone doesn’t know the folly of extending straight lines graphs into the future.

Dnb@lemmy.dbzer0.com on 06 Jul 04:46 collapse

xkcd.com/605

Reminded me of this

And this

xkcd.com/1007

zalgotext@sh.itjust.works on 06 Jul 05:10 collapse

Oof, the alt text on that second one was unexpectedly dark lmao

IAmNorRealTakeYourMeds@lemmy.world on 05 Jul 23:35 next collapse

new moore law dropped

vrighter@discuss.tchncs.de on 06 Jul 06:08 collapse

they are improving at an exponential rate. It’s just that the exponent is less than one.