OpenAI's reported 'superintelligence' breakthrough is so big it nearly destroyed the company, and ChatGPT (www.techradar.com)
from Dadifer@lemmy.world to technology@lemmy.world on 24 Nov 2023 11:35
https://lemmy.world/post/8682717

#technology

threaded - newest

gedaliyah@lemmy.world on 24 Nov 2023 11:48 next collapse

But can it open the pod bay doors?

Enkers@sh.itjust.works on 24 Nov 2023 11:56 next collapse

It can’t do that, Dave.

Bonehead@kbin.social on 24 Nov 2023 12:10 collapse

Dave's not here, man.

reflex@kbin.social on 24 Nov 2023 13:05 collapse

Dave's not here, man.

What about Buster?

random_character_a@lemmy.world on 24 Nov 2023 12:02 collapse

Take your upvote and go watch more artsy 60’s scifi, you brilliant sod.

Nomad@infosec.pub on 24 Nov 2023 19:41 collapse

Is this a dave reference?

DrCake@lemmy.world on 24 Nov 2023 11:48 next collapse

So was it all just a marketing stunt?

otter@lemmy.ca on 24 Nov 2023 11:50 collapse

CEO ousting shenanigans = 📉

Release rumor = 📈

They’re not publicly traded, but I assume public sentiment still has an effect on things (ex. Partnerships, users buying memberships, etc.)

toothbrush@lemmy.blahaj.zone on 24 Nov 2023 11:52 next collapse

just bs. They are trying to come up with an explanation for why altman was fired that is not: we caught him doing lots of illegal stuff.

GONADS125@lemmy.world on 24 Nov 2023 12:38 next collapse

I think it’s a hype move at this point. Like the guy who claimed he believed google’s chat bot was sentient.

I read another article that stated they had a computational breakthrough, in which their program can now carry out basic grade school math. No other model is able to actually carry out math equations, not even basic arithmetic.

This is a significant development, but it’s not like they’re on the cusp of developing superintelligence now. I bet they are taking this small inch towards superintelligence, and hyping it like they’ve just huddled miles forward.

dustyData@lemmy.world on 24 Nov 2023 12:46 next collapse

The thing is, this could actually be a several miles jump. But where they want to go is not the grocery down the road. They are trying to fly to another galaxy. This is more like hyping up that you are going to land on the moon next year, at a time when you just figured out that rubbing two sticks together it makes a fire. Technically it’s truly a leap, but we are so far away still.

GONADS125@lemmy.world on 24 Nov 2023 13:25 collapse

Technically it’s truly a leap, but we are so far away still.

I completely agree and was trying to convey that. Not trying to downplay the significance of the development, but they are far from superintelligence and they’re going to hype it up as much as they can.

Siegfried@lemmy.world on 24 Nov 2023 13:03 next collapse

Is that the chatbot that they had to shutdown cause it wandered a little bit to much in 4chan?

RobotToaster@mander.xyz on 24 Nov 2023 13:38 collapse

That was microsoft’s tay.

Korne127@lemmy.world on 25 Nov 2023 12:33 collapse

The worst part about is it that there have been already two winters in AI development, in the early twothousands and sometimes in the 70/80s? I think? because of exactly this: They always hyped up AI and said they’d solve all the world’s problems in a short time, and when that obviously didn’t happen, people got disappointed in it and pulled funding…

c0mbatbag3l@lemmy.world on 25 Nov 2023 18:08 collapse

Well the models we have now are already useful for things, so it’s unlikely it’ll just disappear now.

We didn’t have the computer technology to make it happen back then, they just didn’t know it at the time.

Korne127@lemmy.world on 30 Nov 2023 17:49 collapse

That’s not my point. We’ve had good AIs and much development in that area of research already 50 years ago. Chess computers started being better than the best humans in the early 2000s. It’s not a particularly new field. But the development and research of artificial intelligence already completely stopped two times and it took over a decade each time to really start research in the field as well.
The reasons why this happened is because of too big promises; even if they succeeded in some things, they promised way too much. If they continue promising way too much in the current AI hype as well, I can see the exact same thing happening again: People getting disappointed and the field getting isolated for another decade.
I’m not saying the current successes will disappear, but that future development might, for a good while, just as it happened back then.

c0mbatbag3l@lemmy.world on 30 Nov 2023 19:18 collapse

None of the previous stabs at AI were more than a parlour trick, modern AI are capable of not only full and natural conversations but have the unique ability to turn that into completing tasks based on how well the human operator can describe the problem and explain the proposed solution.

It’s not always perfect, but it gets close enough for the professional to make use of it by cutting out the research phase of any given project. Or by getting the bulk of the work done without the hours it would have taken to do it. Refining the solution might take ten to fifteen minutes but you don’t have to be a math genius to see the benefits. Plus the models we have now are exploding in niche use-cases. We have image generation, voice generation, code generation, all at near human standards. I’ve had it walk me through how to deploy python scripts via VSC, then I had it walk me through setting up a Git repository, then I asked it to take me through a DnD/Choose your own adventure scenario with specific choices having consequences down the line. It was a little basic but I gave it a preestablished universe and the general premise, it researched the rest on its own and used the data to fill in the gaps in a way I hadn’t even suggested based on what it found of the universe.

That last one isn’t a productive use case, sure. The point is that what we have now isn’t just some one off computer like a chess bot or a Smash Bros CPU set to its highest level, it’s a seed for every future version of machine learning algorithm that will be used to specifically design models for special scenarios. It’s become ingrained in our society now, and it’s unlikely to just disappear like the rest of what you’re describing.

Boiglenoight@lemmy.world on 24 Nov 2023 16:10 collapse

I’m inclined to agree. It just stinks.

teft@startrek.website on 24 Nov 2023 11:52 next collapse

I for one welcome our new robotic super intelligence overlords.

Melt@lemm.ee on 24 Nov 2023 11:55 next collapse

Hope it replaces the most expensive job position: CEO

amir_s89@lemmy.ml on 24 Nov 2023 11:56 next collapse

The whole organization structure & how it functions is just not so smart after all. Have management team considered the Lean Methodology with their business objectives?

FaceDeer@kbin.social on 24 Nov 2023 17:47 collapse

The problem that precipitated all this is that they don't have business objectives. They have a "mission." The board of directors of OpenAI aren't beholden to shareholders, and though the staff mocked their statement that allowing the company to be destroyed “would be consistent with the mission” it's actually true.

amir_s89@lemmy.ml on 24 Nov 2023 19:30 collapse

Appreciate your clerification.

TallonMetroid@lemmy.world on 24 Nov 2023 12:01 next collapse

I’ll believe it when Judgement Day happens and I die in nuclear fire when the wi-fi turns against me.

reflex@kbin.social on 24 Nov 2023 12:01 next collapse

Yawn.

Let me know when we get a real Terminator or Matrix or Space Odyssey situation.

gedaliyah@lemmy.world on 24 Nov 2023 12:41 next collapse

🐈‍⬛

gedaliyah@lemmy.world on 24 Nov 2023 12:41 collapse

🐈‍⬛

Moof_Kenubi@lemmy.world on 24 Nov 2023 15:36 collapse

heck, I’d settle for half a Short Circuit

nicetriangle@kbin.social on 24 Nov 2023 12:03 next collapse

Geez the reporting around this has been ridiculously sensationalist

FlyingSquid@lemmy.world on 24 Nov 2023 12:10 next collapse

You mean OpenAI didn’t just create a superintelligent artificial brain that will surpass all human ability and knowledge and make our species obsolete?

Hyperreality@kbin.social on 24 Nov 2023 12:13 next collapse

You laugh now, but just you wait. If it turns out they've created a hyperintelligent waifu/husbando, this will inevitably lead to plummeting birth rates and the end of human civilisation.

Duke_Nukem_1990@feddit.de on 24 Nov 2023 12:32 next collapse

One can only hope

stifle867@programming.dev on 24 Nov 2023 12:40 next collapse

<img alt="" src="https://programming.dev/pictrs/image/330ae225-11b6-41d3-96e4-18697a0b99f5.png">

FaceDeer@kbin.social on 24 Nov 2023 17:31 collapse

If it's really hyperintelligent it'll realize that human extinction would inconvenience it, and so would Cyrano de Bergerac us into forming healthy relationships with our human soulmates while researching fertility technologies and whatnot.

kescusay@lemmy.world on 24 Nov 2023 12:17 collapse

The funny thing is, last year when ChatGPT was released, people freaked out about the same thing.

Some of it was downright gleeful. Buncha people told me my job (I’m a software developer) was on the chopping block, because ChatGPT could do it all.

Turns out, not so much.

I swear, I think some people really want to see software developers lose their jobs, because they hate what they don’t understand, and they don’t understand what we do.

FlyingSquid@lemmy.world on 24 Nov 2023 12:19 next collapse

Even if ChatGPT gets far in advance of the way it is now in terms of writing code, at the very least you’re still going to need people to go over the code as a redundancy. Who is going to trust an AI so much that they will be willing to risk it making coding errors? I think that the job of at the very least understanding how code works will be safe for a very long time, and I don’t think ChatGPT will get that advanced for a very long time either, if ever.

thisfro@slrpnk.net on 24 Nov 2023 12:23 next collapse

Who is going to trust an AI so much that they won’t risk it making coding errors?

Sadly, too many

FlyingSquid@lemmy.world on 24 Nov 2023 12:24 collapse

Then their companies will go belly-up.

jeena@jemmy.jeena.net on 24 Nov 2023 12:39 next collapse

I don’t believe it. If it’s good enough then they will ship and make money, and those who put people on it will be so slow that they will be just outperformed by those who don’t.

FlyingSquid@lemmy.world on 24 Nov 2023 12:41 collapse

If your code doesn’t work because you rely entirely on an AI to do it, you don’t have a business you can run unless you want to go back to paper and pencil.

jeena@jemmy.jeena.net on 24 Nov 2023 12:45 collapse

If your code doesn’t work because you rely on humans understanding it, you don’t have a business you can run. We already are there where humans have no idea why the computer does this or that decision because it’s so complex especially with all the machine learning and complex training data, etc. let’s not pretend it will get less complex with time.

FlyingSquid@lemmy.world on 24 Nov 2023 12:47 collapse

So your argument is that people will rely on AI entirely without making any redundancies, unlike now where they have more than one human so they can check for these issues because humans make coding errors?

jeena@jemmy.jeena.net on 24 Nov 2023 12:51 next collapse

My argument is that already today no human is able to and checks it when it comes to decision making models like for example if the car should go left or right around a obstacle. And over time we will have less straight forward classical programming doing decisions and more and more models doing decisions with hundreds or thousands of sensor inputs.

FlyingSquid@lemmy.world on 24 Nov 2023 12:53 next collapse

And that means AI code shouldn’t be error-checked?

jeena@jemmy.jeena.net on 24 Nov 2023 13:04 collapse

That means that it right now can not be error checked and it will be even more difficult in the future.

FlyingSquid@lemmy.world on 24 Nov 2023 13:05 next collapse

So you’re saying no code would be worth error checking by a human at all? There is no level of simpler code that an AI could get wrong and would need someone to fix it?

jeena@jemmy.jeena.net on 24 Nov 2023 13:16 collapse

My assumption is that it would even be the other way around, AI checking all human code, especially writing all the tests. So I guess AI would also write tests for it’s own code too.

When it comes to humans, I think they would probably be changing the prompt they give to AI instead of changing the code at the end. You can already see how it’s done with generation of pictures, while theoretically you could take a not so perfect AI picture and use Photoshop to fix it, but most of the time people change the prompt instead and regenerate the picture.

FlyingSquid@lemmy.world on 24 Nov 2023 13:19 next collapse

The generation of pictures is full of fuckups like giving people extra legs, so I’m not sure that’s a very good example.

jeena@jemmy.jeena.net on 24 Nov 2023 13:53 collapse

I’m not sure why you think it’s not a good example. The picture itself is code (PNG, JPEG, etc.) which some AI wrote and no human at the company has checked but it gets delivered to the customer who pays for it. It’s not as good as if a human would do it, but it’s way way way cheaper so you can generate a couple of times until you get what you want. So in this domain humans already are losing jobs to AI even though AI is so bad that it is giving people extra legs, etc.

The same thing happens with hallucinations of ChatGTP, which despite that is still preferable by customers to a human assistant who would summarize articles, etc. with much better quality but very low speed and very expensive.

Code is not much different from summaries of longer texts.

FlyingSquid@lemmy.world on 24 Nov 2023 13:58 collapse

What human has lost their job to AI image creation?

jeena@jemmy.jeena.net on 24 Nov 2023 14:09 collapse

I don’t know any illustrators myself but looking at news it is already happening businessinsider.com/ai-taking-jobs-fears-artists-…

FlyingSquid@lemmy.world on 24 Nov 2023 14:14 collapse

That article says artists are still needed to touch up the AI art, which was my whole point about people who understand code, so I think you just proved my point.

jeena@jemmy.jeena.net on 24 Nov 2023 14:17 collapse

Ok, if your point was that instead of 90 junior programmers and 10 senior programmers before, now the 10 senior programmers will keep their job and instead of touching up the juniors code they touch up the AIs code then yes you’re right, those jobs will be kept for some more time.

FlyingSquid@lemmy.world on 24 Nov 2023 14:25 collapse

My point:

Even if ChatGPT gets far in advance of the way it is now in terms of writing code, at the very least you’re still going to need people to go over the code as a redundancy.

Note I never said anything about the number of people.

Maybe you should read before responding.

jeena@jemmy.jeena.net on 24 Nov 2023 14:27 collapse

Ok, I will do that next time …

Asudox@lemmy.world on 27 Nov 2023 06:48 collapse

Until even the tests are bugged

jeena@jemmy.jeena.net on 27 Nov 2023 07:51 collapse

Have you seen tests written by humans? 😅

Asudox@lemmy.world on 27 Nov 2023 08:40 collapse

Yes.

jeena@jemmy.jeena.net on 24 Nov 2023 13:08 collapse

Or let me rephrase it with the context of the original assumption that if people don’t check the code which AI wrote the company will lose customers because the quality is bad.

Right now there are tons of models out there which no human can understand why they decide this or that, still they bring so much more value that they get shipped even though they make some mistakes. If a company would try to only ship code checked by humans they would not be able to ship the products and would lose their customers to a company which does not check it but does ship.

lemmyvore@feddit.nl on 25 Nov 2023 07:44 collapse

Except we already have fields (like pharma manufacturing) that have to deal with hundreds or thousands of inputs and variables, are automated, and we still manage to fully understand the stack as well as fully check everything.

Hint: when someone tells you they “can’t” check or understand what their software is doing, it’s a scam.

Normally they should be told to go back and figure it out before being allowed to ship any product. If you tried this in any other industry it would be laughable. Even in software it’s outrageous, imagine getting accounting software or even a simple file backup tool that doesn’t work some of the time and nobody can tell you how it works. Yet these companies get a pass putting cars like this on the road.

Enkers@sh.itjust.works on 24 Nov 2023 16:19 collapse

I kinda agree with them. Currently coding already is an abstraction. The average developer has very little idea what machine code their compiler actually produces, and for the most part they don’t need to care about this. Feeding an AI a specification is just a higher level of abstraction.

For now, we’ll need people to check that AI produces code that does what we expect, but I believe at some point we’ll mostly take it for granted that they just do.

thisfro@slrpnk.net on 24 Nov 2023 15:54 collapse

I’d hope so, but it already works for many of them

nicetriangle@kbin.social on 24 Nov 2023 14:02 next collapse

That's a fuckin bleak outcome for a lot of people if the job transition goes from <doing the cool problem solving/creative things> to <cleaning up after the robots' work>

That's like being an artist and being told your job now is simply to fix the shitty hands Midjourney draws. And your job will only last as long as that remains a problem.

FlyingSquid@lemmy.world on 24 Nov 2023 14:03 collapse

Hey, I didn’t say the future would be bright, just that it will still need people familiar with code for the foreseeable future. At least until the Earth heats up so much that the lack of potable water and the unsurvivable high temperatures destroy civilization.

kescusay@lemmy.world on 24 Nov 2023 14:42 next collapse

There’s more to it than that, even. It takes a developer’s level of knowledge to even begin to tell ChatGPT to make something sensible.

Sit an MBA down in front of a ChatGPT window and tell them to make an application. The application has to save state, it has to use the company’s OAuth login system, it has to store data in a PostgreSQL database, and it has to have granular, roles-based access control.

Then watch the MBA struggle because they don’t understand that…

  • Saving state is going to vary depending on the front-end. Are we writing a browser application, a desktop application, or a mobile application? The MBA doesn’t know and doesn’t understand what to ask ChatGPT to do.
  • OAuth is a service running separately to the application, and requires integration steps that the MBA doesn’t know how to do, or ask ChatGPT to do. Even if they figure out what OAuth is, ChatGPT isn’t trained on their particular corporate flavor for integration.
  • They’re actually writing two different applications, a front-end and a back-end. The back-end is going to handle communication with PostgreSQL services. The MBA has no idea what any of that means, let alone know how to ask ChatGPT to produce the right code for separate front-end and back-end features.
  • RBAC is also probably a separate service, requiring separate integration steps. Neither the MBA nor ChatGPT will have any idea what those integration steps are.

The level of knowledge and detail required to make ChatGPT produce something useful on a large scale is beyond an MBA’s skillset. They literally don’t know what they don’t know.

I use an LLM in my job now, and it’s helpful. I can tell it to produce snippets of code for a specific purpose that I know how to describe accurately, and it’ll do it. Saves me time having to do it manually.

But if my company ever decided it didn’t need developers anymore because ChatGPT can do it all, it would collapse inside six months, and everything would be broken due to bad pull requests from non-developers who don’t know how badly they’re fucking up. They’d have to rehire me… And I’d be asking for a lot more money to clean up after the poor MBA who’d been stuck trying to do my job.

FlyingSquid@lemmy.world on 24 Nov 2023 14:44 collapse

Thank you, you explained all of that much better than I could.

kescusay@lemmy.world on 24 Nov 2023 14:57 collapse

You’re welcome! And it occurs to me that the fact that it took a developer to explain all of that is an object lesson in why ChatGPT won’t end software development as a career option - and believe me, I simplified it for a non-developer audience.

archomrade@midwest.social on 24 Nov 2023 17:52 collapse

It isn’t surprising that this is the way we conceptualize the potential impact of AI, but it’s frustrating to see it tossed around as if AI disruption is a forgone conclusion.

AI will start re-defining the problems that code is written to solve long before we get anywhere close to GPT models replacing human workers, and that’s a big enough problem by itself.

It used to be that before code could even be employed to solve a problem, it had to be understood procedurally. That’s increasingly not the case, given that ML is routinely employed to decode things that were previously thought to be too chaotic to be understood, like brain waves and image pixel data. I don’t know why we’re so sure of ourselves that machine learning is just a gimmick and poses no real threat, just because anthropomorphizing it seems silly.

dustyData@lemmy.world on 24 Nov 2023 12:41 next collapse

Your comment reminds me the cesspit of Xitter with the generative AI bros trying to conflate AI with assistive tech. They seriously argued that “artistically impaired” was a genuine disability and that they were entitled to generative AI training sets because it allowed them to draw. It was the most disingenuous argument, that they had a right to steal artists work, and leave them without income, to train their AI because they couldn’t be bothered to rub a pen against some paper.

Valmond@lemmy.mindoki.com on 24 Nov 2023 13:07 next collapse

Artistically impaired, lol that made my day!

RobotToaster@mander.xyz on 24 Nov 2023 13:35 next collapse

I mean, I was literally diagnosed with “clumsy child syndrome” (we call it dyspraxia now) as a kid, in part because I’m artistically impaired.

Jesus_666@feddit.de on 24 Nov 2023 13:39 collapse

Fair point although there is a difference between “can’t make a reasonable drawing with instruction at the level of one’s classmates” and “never progressed beyond very basic drawing skills because you never practiced”.

Peanutbjelly@sopuli.xyz on 24 Nov 2023 14:24 collapse

Hey! Artist here. I love drawing. My hands go numb within minutes and they shake more every year. I appreciate having a tool and medium that allows great artistic control despite these facts.

Now, if you’re really butthurt about the training data you can use adobe’s proprietary model. I for one think it’s good that peasants have an open available tool that isn’t owned by adobe, even if it was trained less proprietarily.

This anger about it reminds me of deviant art artists getting mad at each other for “copying my style”

And the fact that copywrite used to be about the general good, and promotion of creative works.

This world needs new artistic priorities. Pen and paper aren’t losing their place, but new tech will lead to independent artists creating entire movies, games, and holodeck style experiences without looming overhead of whatever art oligarch holds the funding.

dustyData@lemmy.world on 24 Nov 2023 17:00 collapse

Au contraire. Art oligarchs will own every single thing you make with AI. You’re not being liberated, you’re being further imprisoned, and they got you to cheer for the jailers.

ADD: remember, the final goal of the technocrats is not to make more artists. But to remove the artist from the art altogether.

FaceDeer@kbin.social on 24 Nov 2023 17:38 collapse

Art oligarchs will own every single thing you make with AI.

No, where are you getting that from? I'm not even sure how to refute that, it's nonsensical.

There might be some AI services out there that try to use some sort of ToS to "claim" anything you generate using them, but any such service would be radioactive to a serious artist. Just use a different one, or run the AI locally yourself.

Enkers@sh.itjust.works on 24 Nov 2023 13:45 next collapse

As a software developer, I do want to see software developers lose their jobs to AI. This shouldn’t be surprising, as the purpose of a lot of software development is to put other people out of a job via automation, and that’s fundamentally a good thing. The alternative is like wanting a return to preindustrial society. Automation generally raises quality of life.

The real problem is that we still haven’t figured out how to distribute the benefits of society’s automation efforts equitably so that they raise quality of life for everyone.

nicetriangle@kbin.social on 24 Nov 2023 14:06 next collapse

Yeah that would be all fine and well if it meant we're on track for some post-work egalitarian utopia but you and I know that's not at all where this is heading.

Enkers@sh.itjust.works on 24 Nov 2023 14:51 next collapse

Oh, for sure, it’ll definitely further wealth disparity, as automation always seems to in a capitalist system. But that’s a societal problem that we continually have to address, and it spans nearly all fields of human work to varying degrees.

Fortunately, for the most part tech advancements are very hard to control. Progress can be impeded from spreading, but not stopped, and it means the average individual has access to more and more powerful tools.

FaceDeer@kbin.social on 24 Nov 2023 17:41 collapse

Unfortunately based on what I know of history it seems likely that humanity won't ever be on track to build a post-work egalitarian utopia until we've got no other option left. So I support going ahead with this tech because that seems like a good way to force the issue. The transition period will be rough, but better than stagnation IMO.

Eldritch@lemmy.world on 25 Nov 2023 20:10 collapse

We’ve figured it out. They already had a start on it in the 19th and 20th centuries. However, those with the means have spent the last 100 years screaming bloody murder. Dismantling government and any progress that had been made to address it. As well as invading and overthrowing any foreign group that though about opposing them.

Enkers@sh.itjust.works on 25 Nov 2023 20:42 collapse

Glad it’s all figured out. Does that include some sort of viable implementation plan?

Eldritch@lemmy.world on 25 Nov 2023 20:57 collapse

It’s all well planned out. Step one is getting money out of politics. So no, it’s not viable unfortunately.

nicetriangle@kbin.social on 24 Nov 2023 14:00 next collapse

Lotta people have already lost jobs because of it. I know a few personally. People with college educations. We're just getting started with this, it will get worse.

kescusay@lemmy.world on 24 Nov 2023 14:50 collapse

In software development? Not many - and certainly not at smart companies.

ChatGPT is a tool. It goes in the developer toolbox because it’s useful. But it doesn’t replace the developer, any more than a really good screwdriver replaces the construction worker.

More and more, understanding how to use LLMs for software development will be a job requirement, and developers who can’t adapt to that may find themselves unemployed. But most of us will adapt to it fine.

I have. I’m using Copilot these days. It’s great. And the chances of it replacing me are roughly 0%, because it doesn’t actually know anything about our applications, and if told to make code by someone else who doesn’t know anything about them either, it’ll make useless garbage.

nicetriangle@kbin.social on 24 Nov 2023 16:34 collapse

Yeah so your job is a harder to fully replace with AI at this moment than jobs like copywriting, narration, or illustration. Enjoy it while it lasts, because the days are numbered.

And before all those jobs are gone, people using AI tools like your Copilot will be more productive requiring less headcount. At the same time there will still be a lot of people seeking work, but now with fewer jobs there will be downward pressure on wages.

kescusay@lemmy.world on 24 Nov 2023 17:11 collapse

In the early 2000’s, there was all this panic about how these newfangled languages and tools were going to obliterate the developer job market. They were too easy to use! Too simple for non-developers to pick up! Why, you could almost code in plain English now! Developers are doooooooomed!

Instead, demand for developers shot through the roof, because the barrier to entry for developing applications had been lowered enough that adding staff developers to your employee roster became a no-brainer.

Part of the problem is one of precision instructions. We call instructions that are comprehensive, specific, detailed, and precise enough to be turned into programs code, and ChatGPT doesn’t change that. It can only do what you tell it to do.

Maybe someday, a large language model will be so sophisticated, you can say something like, “Write a program to do X, Y, and Z. It uses (address) for authentication, and (other address) for storing data. Here are the credentials to use for each. (Credentials). Your repo is at (address). You deploy the front-end at (address) and the back-end at (address). Your pipelines should be written using (file) as a template.” And maybe what it does with that will truly be able to replace me.

But I genuinely doubt it. I glossed over an enormous amount of detail in that example. If I add it in, what it’ll start looking like is, well, more code.

nicetriangle@kbin.social on 24 Nov 2023 17:20 collapse

This is pretty low level stuff but every dev I've showed this to found it surprising (I hang out on a slack community for designers and devs so it wasn't just like 2 people)

https://www.youtube.com/watch?v=8y7GRYaYYQg

It's just a matter of time.

kescusay@lemmy.world on 24 Nov 2023 19:40 collapse

I didn’t find it surprising, because I do stuff like it every day (except I’m using Copilot integrated into my IDE, not ChatGPT).

What he did was very cool. But he’s a game developer who already knows all the parts he needs and what to ask for, and he still has to do a lot of work by hand. He glossed over it quickly, but there are parts where he had to add code to specific, already-existing blocks of code in his program, and in order to do that, he had to know and understand what the current code was doing.

And throughout the video, he had to know not only what to ask for, but how to ask for it. That takes experience and understanding.

It’s possible that eventually, programming for a lot of people will mean expertise in interacting with large language models + lesser expertise in the actual programming language, but I don’t see that as likely to end programming as we know it. In fact, it might cause a surge in developer demand as the bar to entry lowers again, much like it did in the 2000’s. And there will always be demand for people with a deep understanding of the actual code, because they’ll be necessary for things like performance improvement, bug fixing… and writing the next generation of large language models.

c0mbatbag3l@lemmy.world on 25 Nov 2023 18:06 collapse

Those people were always misinformed.

At some point in the future AI will replace most programmers, because AI will allow senior devs to automate large portions of their codebase. Human devs will act more like QA, fixing the small errors during the automation process.

Either way it’s a tool to by used by you to multiply your efforts, not one to replace you.

DemBoSain@midwest.social on 24 Nov 2023 18:23 next collapse

Superintelligent AI Just Pried the Keyboard from my Cold, Dead Hands

kromem@lemmy.world on 25 Nov 2023 21:31 collapse

Because that’s the version that gets posted and gets clicked on.

A dry technical writeup looking at the name of the project and how it indicates this is a different approach more in line with DeepMind’s work and what that means in the context of doing high school level math is going to be interesting to only a handful of people.

But an article that’s contentious and gets hundreds of comments about how “AI is BS” to “AI is dangerous” all arguing with each other drives engagement.

sentient_loom@sh.itjust.works on 24 Nov 2023 12:16 next collapse

Almost sounds like the whole thing was a performance.

ZILtoid1991@kbin.social on 24 Nov 2023 12:27 next collapse

The "superintelligence" in question: the same old tech, but with a larger context window, which will make it hallucinate a bit less often.

[deleted] on 24 Nov 2023 14:46 next collapse

.

mrnotoriousman@kbin.social on 24 Nov 2023 15:01 collapse

Not really. The headline is garbage but that statement is not even close to accurate unless you know nothing about the actual topic.

Hider9k@lemm.ee on 24 Nov 2023 15:46 next collapse

Explain then pls.

[deleted] on 24 Nov 2023 16:22 collapse

.

ripcord@kbin.social on 24 Nov 2023 18:08 collapse

I believe in this case the breakthrough is the ability to reason out math.

kromem@lemmy.world on 25 Nov 2023 21:27 collapse

Not really.

If it’s name is Q* then it seems likely that it’s a combination of Q learning and A* search, which indicates that this is an approach similar to DeepMind’s AlphaZero as opposed to a transformer based LLM.

In that context, getting it to be able to solve high school level math questions is pretty nuts.

Though the details matter and right now all the articles discussing it are missing those, so we’ll have to wait and see.

Korne127@lemmy.world on 24 Nov 2023 13:20 next collapse

Connecting superintelligence to the board’s recent actions, which Sutskever initially supported, might be a stretch.

Why do you do that in your headline then?

DarkThoughts@kbin.social on 24 Nov 2023 13:58 collapse

Clickbait.

ShaunaTheDead@kbin.social on 24 Nov 2023 14:56 next collapse

Who the hell would have guessed that we'd have to deal with not one but two potentially civilization ending threats in our lifetimes? I want off this crazy ride please!

RiikkaTheIcePrincess@kbin.social on 24 Nov 2023 18:33 next collapse

Why do I keep looking at these threads? The way people talk about this stuff on all sides is so asinine. Nearly every good point is accompanied by missing a big one or just ricocheting off the good one, flying off into space and hitting a fully automated luxury gay space commulist. Hopes, dreams, assumptions, and ignorance all just headbutting each other and getting nowhere.

Oh yeah, I wanted to know what "superintelligence" was and whether I should care. Welp.

Dadifer@lemmy.world on 25 Nov 2023 17:56 collapse

I think the takeaway is that they’re trying to create a LLM that can answer questions that it wasn’t trained on.

satans_crackpipe@lemmy.world on 24 Nov 2023 22:29 next collapse

What are these con artists up to? And why are so many people self replicating the propaganda?

boatswain@infosec.pub on 25 Nov 2023 16:44 collapse

self replicating the propaganda?

You can’t self-replicate anything other than yourself. You replicate things; we use “self-replicating” because it’s shorthand for “thing that replicates itself.”

Tattorack@lemmy.world on 25 Nov 2023 10:52 next collapse

Alright, so the article really doesn’t prove anything, just says OpenAI claims something and then fills it with words.

Let’s be clear here; we don’t even have an AGI. That is to say, artificial general intelligence, a man-made intelligence that is at least as capable and general purpose as Human intelligence.

That would be a intelligence that is self aware and can actually think and understand. Data from Star Trek would be an AGI.

THESE motherfuckers are now claiming they made a breakthrough on potentially creating an SI, a super intelligence. An artificial, man-made intelligence that not only has the self awareness and understanding of an AGI, but is vastly more intelligent than a Human, and likely has awareness that surpasses Human awareness.

I think not.

Enkers@sh.itjust.works on 25 Nov 2023 21:05 collapse

Hahaha. Yeah. :(