First known test dogfight between AI and human pilot carried out, US military says (news.sky.com)
from Aatube@kbin.melroy.org to technology@lemmy.world on 19 Apr 2024 15:54
https://kbin.melroy.org/m/technology@lemmy.world/t/221427

The aircraft flew up to speeds of 1,200mph. DARPA did not reveal which aircraft won the dogfight.

#military #news #technology #usa

threaded - newest

autotldr@lemmings.world on 19 Apr 2024 15:55 next collapse

This is the best summary I could come up with:


In a drill over Edwards Air Force Base, the pair of F-16 fighter jets flew at speeds of up to 1,200mph and got as close as 600 metres during aerial combat, also known as dogfighting.

While in flight, the AI algorithm relies on analysing historical data to make decisions for present and future situations, according to the Defence Advanced Research Projects Agency (DARPA), which carried out the test.

This process is called “machine learning”, and has for years been tested in simulators on the ground, said DARPA, a research and development agency of the US Department of Defense.

In 2020, so-called “AI agents” defeated human pilots in simulations in all five of their match-ups - but the technology needed to be run for real in the air.

Pilots were on board the X-62A in case of emergency, but they didn’t need to revert controls at any point during the test dogfight, which took place in September last year and was announced this week.

"The potential for autonomous air-to-air combat has been imaginable for decades, but the reality has remained a distant dream up until now, said Secretary of the Air Force Frank Kendall.


The original article contains 455 words, the summary contains 193 words. Saved 58%. I’m a bot and I’m open source!

Ghostalmedia@lemmy.world on 19 Apr 2024 16:11 next collapse

We all know how this ends:

www.rottentomatoes.com/m/1146673-1146673-stealth

Aatube@kbin.melroy.org on 19 Apr 2024 16:51 collapse

A bunch of old people bomb the box office?

JimboDHimbo@lemmy.ca on 19 Apr 2024 16:56 collapse

Hey, it was an okay movie for it’s time.

BearOfaTime@lemm.ee on 19 Apr 2024 21:35 collapse

OK is giving it high praise

Blue_Morpho@lemmy.world on 19 Apr 2024 16:13 next collapse

AI will win if not now, then soon. The reason is that even if it is worse than a human, the AI can pull off maneuvers that would black out a human.

Jets are far more powerful than humans are capable of controlling. Flight suits and training can only do so much to keep the pilot from blacking out.

NegativeLookBehind@lemmy.world on 19 Apr 2024 16:19 next collapse

Jets are far more powerful than humans are capable of controlling.

I think the same will eventually be true for AI, especially when you give it weapons

Kyle_The_G@lemmy.world on 19 Apr 2024 16:49 next collapse

I think theres a movie about that

JimboDHimbo@lemmy.ca on 19 Apr 2024 16:55 next collapse

It’s name is Stealth, starring Jamie Foxx.

I can’t believe these idiots went ahead and gave skynet a fucking jet.

pezhore@lemmy.ml on 19 Apr 2024 18:35 collapse

Who knew such a bad movie would be such a good cautionary tale?

intensely_human@lemm.ee on 20 Apr 2024 01:25 collapse

Jamie Foxx

atocci@kbin.social on 19 Apr 2024 17:04 collapse

EDI is a Warplane. EDI must have targets.

intensely_human@lemm.ee on 20 Apr 2024 01:25 collapse

It’s already true for AI. Just observe OpenAI trying to control what their AIs talk about. The mechanisms of control they’re trying to employ are leaky at best.

EdibleFriend@lemmy.world on 19 Apr 2024 16:22 next collapse

Can’t they literally pull turns that would snap the pilots neck?

catsarebadpeople@sh.itjust.works on 19 Apr 2024 17:12 next collapse

Can anyone confirm if AI has a neck?

essteeyou@lemmy.world on 19 Apr 2024 17:22 next collapse

And a mouth ( ͡° ͜ʖ ͡°)

Blue_Morpho@lemmy.world on 19 Apr 2024 18:31 next collapse

But can they scream?

intensely_human@lemm.ee on 20 Apr 2024 01:26 collapse

Not very loud when I’m using it

catsarebadpeople@sh.itjust.works on 19 Apr 2024 19:02 collapse

ಠ_ಠ

Aatube@kbin.melroy.org on 19 Apr 2024 17:59 next collapse

But the supervising wingmen do.

kambusha@lemmy.world on 19 Apr 2024 21:17 collapse

My neck. My back. Lick my inputs & my headphone-jack.

shortwavesurfer@monero.town on 20 Apr 2024 03:54 collapse

You deserve an updoot for this. Enjoy.

Poem_for_your_sprog@lemmy.world on 19 Apr 2024 17:41 next collapse

Maybe if you were sitting sideways in the cockpit and did it very abruptly with the flight control computer disabled (only a few jets can even disable it). It’s the sustained G loading that makes you black out or red out.

gravitas_deficiency@sh.itjust.works on 19 Apr 2024 23:36 collapse

A skilled and fit pilot can pull ~9G in a Viper for about 30s.

A computer can pull ~9G for as long has the plane has the speed to pull that hard, or it can pull as hard as it can until the plane snaps in half, because computers don’t suffer from g-LOC.

BrightCandle@lemmy.world on 19 Apr 2024 17:23 next collapse

Not so much f16s but the more modern planes can do 16G where the pilot can’t really do more than 9G. But once unshackled from a pilot a lot of instrument weight and pilot survival can be stripped from a plane design and the airframe built to withstand much more, with titanium airframes I see no reason we can’t make planes do sustained unstable turns in excess of 20G.

intensely_human@lemm.ee on 20 Apr 2024 01:27 collapse

We might have to start submerging pilots in breathable fluid.

Buelldozer@lemmy.today on 19 Apr 2024 17:45 next collapse

AI will win if not now, then soon.

This article didn’t mention it but the AI pilot did win at least one of the engagements during this testing run.

tal@lemmy.today on 19 Apr 2024 20:39 collapse

Not that that isn’t interesting, but I’d jump in and insert a major caution here.

I don’t know what is being done here, but a lot of the time, wargaming and/or military exercises are presented in the media as being an evaluation of which side/equipment/country is better in a “who would win” evaluation.

I’ve seen several prominent folks familiar with these warn about misinterpreting these, and I’d echo that now.

That is often not the purpose of actual exercises or wargames. They may be used to test various theories, and may place highly unlikely constraints on one side that favor it or the other.

So if someone says “the US fought China in a series of wargames in the Taiwan Strait and the US/China won in N wargames”, that may or may not be because the wargame planners were trying to find out who is likely to win an actual war, and may or may not have much to to with the expectations the planners have of a win in a typical scenario. They might be trying to find out what would happen in a particular scenario that they are working on and how to plan for that scenario. They may have structured things in a way that are not representative of what they expect to likely come up.

To pull up an example, here’s a fleet exercise that the US ran against a simulated German fleet between World War I and II:

en.wikipedia.org/wiki/Fleet_problem

Fleet Problem III and Grand Joint Army-Navy Exercise No. 2

During Fleet Problem III, the Scouting Force, designated the “Black Force,” transited from its homeport in the Chesapeake Bay towards the Panama Canal from the Caribbean side. Once in the Caribbean, the naval forces involved in Fleet Problem III joined with the 15th Naval District and the Army’s Panama Division in a larger joint exercise.[9] The Blue force defended the canal from an attack from the Caribbean by the Black force, operating from an advance base in the Azores. This portion of the exercise also aimed to practice amphibious landing techniques and transiting a fleet rapidly through the Panama Canal from the Pacific side.[10]

Black Fleet’s intelligence officers simulated a number of sabotage operations during the course of Fleet Problem III. On January 14, Lieutenant Hamilton Bryan, Scouting Force’s Intelligence Officer, personally landed in Panama with a small boat. Posing as a journalist, he entered the Panama Canal Zone. There, he “detonated” a series of simulated bombs in the Gatun Locks, control station, and fuel depot, along with simulating sabotaging power lines and communications cables throughout the 16th and 17th, before escaping to his fleet on a sailboat.

On the 15th, one of Bryan’s junior officers, Ensign Thomas Hederman, also snuck ashore to the Miraflores Locks. He learned the Blue Fleet’s schedule of passage through the Canal from locals, and prepared to board USS California (BB-44), but turned back when he spotted classmates from the United States Naval Academy - who would have recognized and questioned him - on deck. Instead, he boarded USS New York (BB-34), the next ship in line, disguised as an enlisted sailor. After hiding overnight, he emerged early on the morning of the 17th, bluffed his way into the magazine of the No. 3 turret, and simulated blowing up a suicide bomb - just as the battleship was passing through the Culebra Cut, the narrowest portion of the Panama Canal. This “sank” New York, and blocked the Canal, leading the exercise arbiters to rule a defeat of the Blue Force and end that year’s Grand Joint Army-Navy Exercise.[11][10] Fleet Problem III was also the first which USS Langley (CV-1) took part in, replacing some of the simulated aircraft carriers used in Fleet Problem I.[12]

That may be a perfectly reasonable way of identifying potential weaknesses in Panama Canal transit, but the planners may not have been aiming for the overall goal of evaluating whether, in the interwar period, Germany or the US would likely win in an overall war. Saying that the Black Fleet defeated the Blue Fleet in terms of the rules of the exercise doesn’t mean that Germany would necessarily win an overall war; evaluating that isn’t the purpose of the exercise. If, afterwards, an article says “US wargames show that interwar Germany would most likely defeat the US in a war”, that may not be very accurate.

For the case OP is seeing, it may not even be the case that the exercise planners expect it to be likely for two warplanes to get within dogfighting range. We also do not know what, if any, constraints were placed on either side.

Gigan@lemmy.world on 19 Apr 2024 18:14 next collapse

Jets are far more powerful than humans are capable of controlling. Flight suits and training can only do so much to keep the pilot from blacking out.

Can they be piloted remotely? Or would that be too dangerous with latency

Aatube@kbin.melroy.org on 19 Apr 2024 18:17 next collapse

You're better off with drones

Gigan@lemmy.world on 19 Apr 2024 18:28 next collapse

But can a drone fly like an F-16?

Aatube@kbin.melroy.org on 19 Apr 2024 18:35 next collapse
Blue_Morpho@lemmy.world on 19 Apr 2024 21:14 collapse

If an f16 is remote controlled, it’s technically a drone.

grue@lemmy.world on 19 Apr 2024 19:25 collapse

What’s the difference? A remotely or AI-piloted fighter jet is just a big drone.

intensely_human@lemm.ee on 20 Apr 2024 01:30 next collapse

The only thing that makes it bigger is the cockpit. There’s no difference.

Aatube@kbin.melroy.org on 20 Apr 2024 02:15 collapse

Drones are designed without cockpits. Retrofitting remote-control into an F-16 does not seem like the best choice to me.

freeman@sh.itjust.works on 20 Apr 2024 07:30 collapse

Retrofitting F-16s to become drones (whether rc or ai-controlled) as well as designing a variant ditching human support for weight and monetary gains is the rational choice as long as non stealth aircraft are viable. In that case you’d stick to F-35s.

It makes no sense to waste billions worth of perfectly capable and proven airframes, engines and avionics. Any future drone that will have at least the same level of capabilities as an f-16 will cost practically cost the same. At the cost of high performance aircraft life support does not add that much cost to a plane, pilot costs (and availability) are a much bigger issue.

psud@lemmy.world on 19 Apr 2024 18:28 next collapse

Yes they can. Before AI the US was expecting to move to remote piloted jets

intensely_human@lemm.ee on 20 Apr 2024 01:29 collapse

Expecting to? We’ve been using remote piloted jets for twenty years.

psud@lemmy.world on 20 Apr 2024 06:43 collapse

That’s not the case yet for fighters, just things like predator drones and global hawk

So really just surveillance and delivery of a couple of light air to surface missiles, most reported on for assassinations

intensely_human@lemm.ee on 20 Apr 2024 01:29 collapse

Latency, signal interference, and limited human intelligence are all limiting factors in that strategy.

If the enemy interferes with any of those, the enemy wins.

This was is already being fought with autonomous drones. By the end of it, the robots will be unrecognizable to us now.

GBU_28@lemm.ee on 19 Apr 2024 18:23 next collapse

Plus the ai has no risk, outside of basic operation.

Humans have an inherent survival instinct to which drones can just say “lol send the next one I’m dying cya”

Aatube@kbin.melroy.org on 19 Apr 2024 18:24 next collapse

Jets are a lot more expensive. What's at risk is all these resources for the jet going down the drain.

GBU_28@lemm.ee on 19 Apr 2024 18:44 next collapse

Huh? Jets are far more replaceable than a human operator who takes years of training and has “needs”.

Ya know unless your military is running on cold war fumes or something and you can’t afford to build an airframe you already have in production

diffusive@lemmy.world on 19 Apr 2024 19:06 collapse

Training a combat pilot used to cost (in early 2000, not sure now) 10M€ for a NATO member.

Find me a modern jet that costs so little. Regardless of what politicians say, human life has a price… and it is waaaay below a jet (even including the training)

grue@lemmy.world on 19 Apr 2024 19:24 next collapse

Yeah, but procurement of a combat pilot has about a two-decade lead time. You can build more jets a lot quicker (potentially even including the R&D phase).

intensely_human@lemm.ee on 20 Apr 2024 01:24 collapse

Also as this war expands to become planet-wide, industrial output of drones will expand many orders of magnitude.

GBU_28@lemm.ee on 19 Apr 2024 20:49 collapse

It’s not just money. It’s time, public perception, quantity trainers, quantity student seats etc

A drone is ready the moment it comes off the assembly line, is flashed with software, and tested.

everyone_said@lemmy.world on 19 Apr 2024 21:09 next collapse

I’d imagine they’d evetually design a jet purpose built for an AI that would be a lot cheaper than a human-oriented one. Removing the need for a cockpit with seats, displays, controls, oxygen, etc would surely reduce cost. It would also open the door for innovations in air-frame design previously impossible.

intensely_human@lemm.ee on 20 Apr 2024 01:23 next collapse

We keep talking like we’re discussing the future, but autonomous drones are already fighting in the skies of Ukraine.

Begun, the drone wars have.

Bgugi@lemmy.world on 21 Apr 2024 03:24 collapse

Jets are in many ways expensive because they can’t be expendible. They also make an bunch of compromises to accommodate keeping a human alive.

For the cost of a single f22, you could put up 60 Valkyries. I think I know which side I would bet on.

intensely_human@lemm.ee on 20 Apr 2024 01:22 collapse

To fight optimally, AI needs to have a survival instinct too.

Evolution didn’t settle on “protect my life at all costs” as our default instinct, simply by chance. It did so because it’s the best strategy in a hostile environment.

maynarkh@feddit.nl on 20 Apr 2024 09:48 next collapse

It’s the best strategy because it takes decades to make a fully functional human, and you need humans to make more humans, plus there is the issue of genetically sustainable population sizes, etc. A fully functional aeroplane can be made much quicker, in a factory that can spit out several of them in a day. They are more expendable.

Turun@feddit.de on 21 Apr 2024 12:12 collapse

Only if the goal is reproduction. You need to survive to reproduce.

If the goal is maximum damage for the least amount of economic cost then a suicide (anthropomorphizing the drone here) can very much make sense.

No one would argue that a sword is better than guns or bombs, because you still have the sword after attacking.

circuscritic@lemmy.ca on 19 Apr 2024 19:18 next collapse

Maneuverability is much less of a factor now as BVR engagements and stealth have taken over.

But, yeah, in general a pilot that isn’t subject to physical constraints can absolutely out maneuver a human by a wide margin.

The future generation will resemble a Protoss Carrier sans the blimp appearance. Human controllers in 5th and 6th gen airframes who direct multiple AI wingman, or AI swarms.

intensely_human@lemm.ee on 20 Apr 2024 01:20 collapse

Here comes the juice!

inb4_FoundTheVegan@lemmy.world on 19 Apr 2024 16:40 next collapse

One step closer to machine domination.

Like, not even in a joking sense. Ukraine is using a ton of drones, the future of physical warfare will simply be a test resources and production.

I’m honestly not sure if this will be good or bad in the longterm. Absolutely saving any amount of human life is a good thing, but when that is no longer a significant factor, I wonder if we will go to (and stay at) war for more trivial reasons.

xor@lemmy.blahaj.zone on 20 Apr 2024 11:08 collapse

I’m hoping that the sheer cost of executing that sort of war will continue to be a prohibiting factor like it is today

ininewcrow@lemmy.ca on 19 Apr 2024 16:45 next collapse

AI has a already won in these confrontations

Surface To Air missile made human piloted aircraft obsolete.

All that’s needed now are a bunch of missiles, plug into an AI program and let it run by itself.

Why would militaries invest in a billion dollar aircraft piloted by a highly trained aircraft pilot with years of training that cost millions of dollars that is probably paid millions over many years … when the pilot and his aircraft can be shot down by a $100,000 missile. If you can’t do it with one missile, send three, four or ten, it’s still cheaper than matching them with an aircraft and pilot.

Instead of investing in expensive aircraft and pilots, all a defending country can do is just spend the same amount of money and surround their country with anti aircraft missiles controlled by AI systems.

Aatube@kbin.melroy.org on 19 Apr 2024 16:47 next collapse

Are you not aware of stealth aircraft and evasive maneuvers? Not to mention that missiles won't necessarily down aircraft.

chakan2@lemmy.world on 19 Apr 2024 17:54 collapse

How do you deploy manned fighters against an aircraft you can’t detect?

Edit: As to not downing the aircraft. That’s irrelevant. It wouldn’t matter if it’s an air to air missile or ground to air…you just fire more.

I can buy thousands of rockets for the cost of a single F22 or F16.

Buelldozer@lemmy.today on 19 Apr 2024 18:46 collapse

How do you deploy manned fighters against an aircraft you can’t detect?

Detect and target aren’t the same thing. There’s various Air Defense platforms that can detect stealth air craft but they lack the resolution necessary to target them. For targetting the launch platform has to be a lot closer.

I can buy thousands of rockets for the cost of a single F22 or F16.

Annnd were back to the Air Defense platforms being hideously expensive. Literally no one can afford enough of them to cover more than a tiny fraction of their air space.

Forget “thousands” of missiles any country larger than a Lichtenstein would to need to buy millions of them along with enough Ground Detection and Launch Stations to cover their entire border. Utterly and totally unaffordable.

Aatube@kbin.melroy.org on 19 Apr 2024 19:28 next collapse

Hurrah Liechtenstein, safest country on planet!

chakan2@lemmy.world on 20 Apr 2024 01:09 collapse

You understand you need all that equipment to get a manned fighter close enough to engage. At that point firing a barrage of smart missiles is still cheaper.

Aatube@kbin.melroy.org on 20 Apr 2024 17:52 collapse

And what equipment would that be?

chakan2@lemmy.world on 21 Apr 2024 14:42 collapse

The big cost would be satellite and radar coverage. We have that already.

GluWu@lemm.ee on 19 Apr 2024 17:42 next collapse

There’s also the extreme bottleneck of using a meat bag that dies if it maneuvers too fast or for too long. The limitations of aircraft are not their own strength or materials.

Buelldozer@lemmy.today on 19 Apr 2024 17:50 collapse

Why would militaries invest in a billion dollar aircraft piloted by a highly trained aircraft pilot with years of training that cost millions of dollars that is probably paid millions over many years … when the pilot and his aircraft can be shot down by a $100,000 missile.

  1. Force projection.
  2. It ain’t that easy to shoot down stealth aircraft.
  3. Missiles that can successfully shoot down stealth aircraft cost several million dollars each.
  4. Ground launch systems that can target and engage stealth aircraft, like the US Patriot System, are so horrifically expensive that no nation can afford enough of them to cover more than a fraction of its airspace. That means you need aircraft capable of engaging incoming enemy targets.

In short you are hideously naive.

flying_monkies@kbin.social on 19 Apr 2024 18:56 collapse

It ain’t that easy to shoot down stealth aircraft.

Hell, it's not easy to shoot down non-stealth aircraft

The pilot evaded all six missiles while not being able to deploy any chaff/flare countermeasures.

PipedLinkBot@feddit.rocks on 19 Apr 2024 18:56 collapse

Here is an alternative Piped link(s):

Hell, it’s not easy to shoot down non-stealth aircraft

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I’m open-source; check me out at GitHub.

JoMiran@lemmy.ml on 19 Apr 2024 17:02 next collapse

<img alt="" src="https://i.imgflip.com/19tgd2.jpg">

ivanafterall@kbin.social on 19 Apr 2024 17:15 next collapse

I'm not an AI expert, but I think it's got the plane turned the wrong way around.

A_A@lemmy.world on 19 Apr 2024 17:35 collapse

it’s the ground that’s upside down because it happened on the other side of the globe 😋.

Buelldozer@lemmy.today on 19 Apr 2024 17:42 next collapse

The F-16 being flown by the AI has 3-D Thrust Vectoring.

It’s an absolute beast of an aircraft.

SatansMaggotyCumFart@lemmy.world on 19 Apr 2024 19:17 next collapse

So does my penis and it is not a beast of a penis.

Cocodapuf@lemmy.world on 20 Apr 2024 05:35 collapse

That seems like cheating if it’s being put up against a standard f16…

EndOfLine@lemmy.world on 19 Apr 2024 18:08 next collapse

In 2020, so-called “AI agents” defeated human pilots in simulations in all five of their match-ups - but the technology needed to be run for real in the air.

It did not reveal which aircraft won the dogfight.

I’m gonna guess the AI won.

Dkarma@lemmy.world on 19 Apr 2024 18:15 next collapse

You think aliens are actually piloting the craft that come all this way?

You’re never gonna see an alien body because they aren’t here.

Bet we’ve got a drone or two of theirs tho.

Aatube@kbin.melroy.org on 19 Apr 2024 18:18 collapse

I am very confused.

psud@lemmy.world on 19 Apr 2024 18:26 next collapse

Dkarma’s comment requires context. They think aliens have visited Earth. They presume people who don’t agree with them expect that if aliens had visited Earth, some would have been shot down and alien bodies would have been recovered

BearOfaTime@lemm.ee on 19 Apr 2024 21:25 collapse

I still don’t get it.

What’s that have to do with this post?

psud@lemmy.world on 19 Apr 2024 22:25 collapse

It has nothing to do with the post. I believe the post inspired dkarma to work out that brand new argument that there are no alien bodies (at area 51 probably) because the aliens would use drones

bradorsomething@ttrpg.network on 20 Apr 2024 04:54 collapse

When you want to hammer alien conspiracies, everything looks like a nalien.

xor@lemmy.blahaj.zone on 20 Apr 2024 11:06 collapse

Upvoted exclusively for that 10/10 pun

unreasonabro@lemmy.world on 19 Apr 2024 18:31 collapse

it’s ok, almost nobody else is paying attention to what really matters either

intensely_human@lemm.ee on 20 Apr 2024 01:14 collapse

Whether aliens are visiting us matters just about as much as whether tanks are rolling into the village of uncontacted tribes.

Our tactical disadvantage against alien technology is zero, so they have zero priority as targets.

Our best bet is to make friends, converse with them. But they are obviously not interested in talking. So our only option is to pray they continue to let us exist. So far they seem to be.

intensely_human@lemm.ee on 20 Apr 2024 01:14 collapse

Sorry, technical advantage is zero. Not disadvantage. We have zero capability to counter alien threats.

Aatube@kbin.melroy.org on 20 Apr 2024 02:13 collapse

You realize that you may edit? :)

intensely_human@lemm.ee on 20 Apr 2024 08:18 collapse

I like immutable records. If the contents of the edit were transparent (are they? in the mod log or something?), I’d have no problem with it. Like if the UI showed the final state but I had the whole log of creates and updates available to inspect like wikipedia, that would be cool

intensely_human@lemm.ee on 20 Apr 2024 08:19 collapse

Also, lemmy doesn’t seem to support strikethrough, so even if people trust me not to be sneaky, I can’t easily show what words were struck and added

Glytch@lemmy.world on 19 Apr 2024 20:02 next collapse

I was actually assuming the opposite, because if the AI won they’d want to brag about it.

intensely_human@lemm.ee on 20 Apr 2024 01:11 next collapse

Bragging just means more money flowing to enemies’ research programs. When a fight is inevitable you want to appear as weak as possible to prevent your enemy from taking it seriously.

Cethin@lemmy.zip on 20 Apr 2024 05:52 collapse

No way we give up that information for free. Either way it went, the knowledge of it cost a lot to gain and is useful. If it failed you want your enemy wasting money on it. If it succeeded you want your enemy not investing in it.

Harbinger01173430@lemmy.world on 20 Apr 2024 00:04 next collapse

Hahaha how the fuck is AI going to win in air-to-air combat if we completely delete them when playing Ace Combat in the highest difficulty?

Seethe, AI tech bros.

AdamEatsAss@lemmy.world on 20 Apr 2024 00:40 next collapse

I know right. Aimbots have been around for years.

Harbinger01173430@lemmy.world on 20 Apr 2024 00:45 collapse

Tech bros are probably patting themselves in the back for ‘inventing’ difficulty Normal++

SendMePhotos@lemmy.world on 20 Apr 2024 02:12 collapse

So instead of Ai vs human fight… It’s like human vs time to hack?

Harbinger01173430@lemmy.world on 20 Apr 2024 04:02 collapse

Yeah, time to add aimbots in the human player fighter jet to make the tech bros seethe in anger even more.

SendMePhotos@lemmy.world on 20 Apr 2024 20:00 collapse

We’ll just start using airburst EMPs

Everythingispenguins@lemmy.world on 20 Apr 2024 00:08 collapse

Or it had a fatal error and crashed the plane

unreasonabro@lemmy.world on 19 Apr 2024 18:29 next collapse

giving AI military training is “responsible”, is it? Oh good, I’m glad training software to kill is going “responsibly”, that’s good to know. Kinda seems like the way a republican uses words - backwards, in opposition to their actual meaning, but hey, fuck the entire world, right?

tal@lemmy.today on 19 Apr 2024 20:09 next collapse

If you want some sort of arms control agreement for AI, you’re going to be faced with the problem of verification that countries are complying.

My guess is that that’s probably very difficult to do. All you need is a datacenter somewhere and someone with expertise.

And if an arms control agreement doesn’t exist, then a country not developing a promising technology just disadvantages that country.

intensely_human@lemm.ee on 20 Apr 2024 01:20 collapse

And if an arms control agreement does exist, it’s just a trap for those naive enough to think such things work.

Putin got us to avoid prepping for a Ukraine invasion simply by repeating that he wasn’t going to invade. And right up until the very moment it happened, the dominant conversation still was not based on the premise that he was going to.

The whole concept of doublespeak works because humans have a powerful compulsion to simply believe what others say. Even if we know their actions and their words are in conflict, we have an extremely hard time following our observations of their actions, and ignoring their words.

It’s like the Stroop task, but with other humans’ behavior instead of ink colors.

intensely_human@lemm.ee on 20 Apr 2024 01:16 collapse

Conservatives tend to be those who, by experience, have been forced out of the notion that the base of existence is not war.

It’s an illusion which can only be maintained when others are facing the war.

Humans tend to remain in the comfortable illusion until they are forced out of it, usually by an encounter with a psychopath victimizing them, or an actual war.

sanae@lemmy.world on 19 Apr 2024 19:01 next collapse

Ace Combat 7 IRL

werefreeatlast@lemmy.world on 19 Apr 2024 19:42 next collapse

Garodotooo! Gundamapailotu! Uhh, ahhh! Gundama!

homesweethomeMrL@lemmy.world on 19 Apr 2024 19:45 next collapse

Planes have been able to be remote-controlled since the 80s.

Since the 80s.

tal@lemmy.today on 19 Apr 2024 20:03 collapse

I mean, yeah, but the significant bit here isn’t remote control by a human, but that there isn’t a human running things.

TheControlled@lemmy.world on 19 Apr 2024 20:26 next collapse

Imagine the pressure on the pilot from their buddies… You would never live down losing

6daemonbag@lemmy.dbzer0.com on 19 Apr 2024 21:04 next collapse

Is Sharon Apple somehow involved? Can we get a list of virtual idols with suspicious behavior?

chatokun@lemmy.dbzer0.com on 19 Apr 2024 21:17 next collapse

Instantly my thought too.

FenrirIII@lemmy.world on 20 Apr 2024 00:02 collapse

That’s a Macross reference?

KISSmyOSFeddit@lemmy.world on 19 Apr 2024 21:13 next collapse

Are dogfights even still a thing?
I remember playing an F15 simulator 20 years ago where “dogfighting” already meant clicking on a radar blip 100 miles away, then doing something else while your missile killed the target.

VindictiveJudge@lemmy.world on 19 Apr 2024 22:48 next collapse

‘Dogfighting’ mostly just means air-to-air combat now. They do still make fighter jets that have guns or can mount guns, but I think they’re primarily intended for surface targets rather than air targets.

intensely_human@lemm.ee on 20 Apr 2024 01:09 collapse

I always thought it would be a game change to have guns that could point different directions.

Cocodapuf@lemmy.world on 20 Apr 2024 05:31 next collapse

Many aircraft guns do that. It’s also usually automatic, look a direction and the gun points where you look.

Cethin@lemmy.zip on 20 Apr 2024 05:50 next collapse

I honestly wouldn’t be that surprised if an AI powered fighter jet got point defence systems installed. It could react to put an incoming missile directly in the path of the point defense and possibly shoot it before it hits. With that said, I don’t know how useful it’d be. If it’s coming right at you the shrapnel is still on its way. Maybe it can react and plan in such a way to avoid it. I guess it depends on the relative speed and direction of the incoming missile.

intensely_human@lemm.ee on 20 Apr 2024 08:21 collapse

I meant to just shoot the other plane without having to point your plane straight at them

Cethin@lemmy.zip on 20 Apr 2024 09:28 collapse

Most dogfights are done by missiles these days. You lock on, fire, and forget. A ballistic projectile wouldn’t be useful at those ranges. I guess if AI fighter jets change progression to bring things back to close-in combat something like that may be useful. I don’t predict it’d happen, but there’s a chance. Missiles can just pack so much more thrust-per-second into a small package than a jet can, but AI jets could be a lot more agile and essentially a missile firing missile, or similar, letting them close distance better whole avoiding incoming ordinance.

echodot@feddit.uk on 20 Apr 2024 10:47 collapse

So turrets?

They had that back in the first world war.

freeman@sh.itjust.works on 20 Apr 2024 06:23 collapse

Well if both sides get working stealth dogfights are going to become more common.

But the US seems to estimate it’s adversaries do not have such capability at the moment since it’s ordering new F-15s with the major change being air to air missile capacity.

Missiles also did not have 100 miles range 20 years ago. That’s without considering actually detecting and tracking the target.

echodot@feddit.uk on 20 Apr 2024 10:46 collapse

Missiles also did not have 100 miles range 20 years ago.

Somewhat missing then point there I feel.

They are right, I was thinking the exact same thing when I read the headline, aircraft don’t really engage in dog fights anymore. It’s all missiles and long-range combat. I don’t think any modern war would involve aircraft shooting at each other with bullets.

freeman@sh.itjust.works on 20 Apr 2024 12:45 collapse

No not really, my point is that people have a distorted and exaggerated view of BVR. 100 miles was beyond even the max range of common missiles and even with modern missiles like meteor it’s completely unrealistic to fire at that kind of range. Provided that you have detected and are able to track the target at that range.

I don’t know if modern planes will have to resort to guns but WVR dogfights with IR missiles are more likely than destroying F-35s at BVR ranges.

BlackNo1@lemmy.world on 20 Apr 2024 01:00 next collapse

i hate this world

intensely_human@lemm.ee on 20 Apr 2024 01:09 next collapse

We all know which aircraft won the fight.

Those of us who play video games do at least. All the AI difficulty settings are arbitrary. You give the bot the ability to use its full capability, and the game is unplayable.

Jimmycakes@lemmy.world on 20 Apr 2024 01:32 next collapse

Plus they had humans on board the AI jet. I imagine it could pull some crazy insane Gs without the human pushing the engineering to the red line.

Cethin@lemmy.zip on 20 Apr 2024 05:45 next collapse

For sure without humans the AI probably wins, assuming the instruments are good. This wasn’t without humans, but it probably still wins.

I’m fairly certain most dogfights happen on instruments only at this point, so I don’t see a chance the human won. The AI can react faster and more aggressively. It can also almost perfectly match a G-load profile limit (which could be much higher without humans on board) where a human needs to stay a little under to not do damage.

This is all assuming the data it was given was good and comprehensive, which I’m sure it was. It also likely trained in a simulation a lot too. This is one of those things AI is great for. Anything that requires doing something new and unique it can’t handle, but if it just requires executing an output based on inputs, that’s a perfect use case.

intensely_human@lemm.ee on 20 Apr 2024 08:23 next collapse

What if we invent artificial gravity just so we can simulate pilot orientation and g forces while they sit still in a simulator?

echodot@feddit.uk on 20 Apr 2024 10:40 next collapse

Whataboutism taken to its extreme there.

Hell, what if we invented warp drive that allowed us to teleport bombs directly into our enemies headquarters?

magnusrufus@lemmy.world on 20 Apr 2024 12:16 collapse

Nonsense hypotheticals are not whataboutism.

intensely_human@lemm.ee on 20 Apr 2024 17:59 collapse

Thank you

Aatube@kbin.melroy.org on 20 Apr 2024 14:51 collapse

We already have g-force simulation.

intensely_human@lemm.ee on 20 Apr 2024 17:59 collapse

No we have g-force production. Until we release those electrogravitics from the top secret labs we can’t actually simulate g forces.

Aatube@kbin.melroy.org on 20 Apr 2024 20:04 collapse

Electrogravitics seem like a conspiracy theory. Unless they've been around as long as human centrifuges, which DO simulate g-forces, I doubt that they'd be more economical even if they do exist.

intensely_human@lemm.ee on 21 Apr 2024 10:42 next collapse

No centrifuges create g-forces. The forces you feel in a centrifuge are actual g-forces.

Aatube@kbin.melroy.org on 21 Apr 2024 12:39 collapse

Why do we need "authentic" g-forces to be "created"? As you've said, people already feel g-forces in centrifuges.

Natanael@slrpnk.net on 21 Apr 2024 13:58 collapse

There is a connection between gravity and electromagnetics, but it’s mostly through the stress-energy tensor giving photons momentum (and thus gravitational pull) but to use an EM field to measurable gravity you need absolutely insane amounts of energy.

You essentially need the literal inverse of a supermassive nuclear explosion (almost like a small star), because the gravitational effect of energy is equivalent to the gravitational effect of the mass which it would form if bound, and given E=mc^2 and the fact that nuclear bombs are small enough to barely have measurable gravity then the math means you need truly insane amounts of energy. (unless somebody can figure out a cheat to create directional pull with much less energy, but I strongly doubt it)

It’s more plausible that somebody would be able to scale up “optical tweezers” to move large masses (directly depositing momentum of the energy field on an object) because that no longer involves the E=mc^2 equation, but it would be even more complicated by a HUGE factor than building the type of large supercooled electromagnets which already can make humans hover (due to water in the body being diamagnetic)

echodot@feddit.uk on 20 Apr 2024 10:43 collapse

I don’t know, one camera lead falls out and it’s all over for the AI. The human still is going to be more adaptable than an AI and always will be until we have full true AGI.

Having said that if we ever do have AGI I 100% believe the US military would be stupid enough to put it in a combat aircraft.

Bgugi@lemmy.world on 20 Apr 2024 22:52 next collapse

What if the pilot has a stroke?

intensely_human@lemm.ee on 21 Apr 2024 10:41 collapse

Yeah but a camera lead falling out isn’t normal

calcopiritus@lemmy.world on 20 Apr 2024 08:36 collapse

In video games the AI have access to all the data in the game. In real life both the human and AI have access to the same (maybe imprecise) sensor data. There are also physical limitations in the real world. I don’t think it’s the same scenario.

BluesF@lemmy.world on 20 Apr 2024 18:50 next collapse

Not exactly, AI would be able to interpret sensor data in a more complete and thorough way. A person can only take in so much information at once - AI not so limited.

calcopiritus@lemmy.world on 20 Apr 2024 20:20 collapse

Don’t get me wrong. Humans have many limitations that AI don’t in this scenario. I’m not saying that a human would do better. For example, as others have stated, an AI doesn’t suffer from G forces like a human does. AI also reads the raw sensor data instead of a screen.

All I’m saying that this case is not the same as a videogame.

intensely_human@lemm.ee on 21 Apr 2024 10:40 collapse

Video games can model point of view and limit AI to what they can legitimately see, while still taking the governor chip off their aiming and reaction time performance.

NeatNit@discuss.tchncs.de on 21 Apr 2024 12:38 collapse

While true, I wonder how many games actually do this.

VirtualOdour@sh.itjust.works on 21 Apr 2024 14:49 collapse

Ai can balance a physical triple pendulum and move between positions fluidly just using vision alone, a human has no chance at coming close.

We’re genuinely in the sci-fi robots are better than humans phase of history, by 2030 you’ll be used to seeing impressive things done by robots like dude perfect videos with people setting up crazy challenges like ‘I got my robot to throw THIS egg through THIS obstacle course and you’ll never belive how it did it!’

NeatNit@discuss.tchncs.de on 21 Apr 2024 12:38 collapse

I think even the imperfect sensor data is enough to beat a human. My main argument for why self-driving cars will eventually be objectively safer than the best human drivers (no comment about whether that point has already done) is this:

A human can only look at one thing at a time. Compared to a computer, we see allow, think slow, react show, move slow. A computer can look in all directions all the time, and react to danger coming from any of those directions faster than a human driver would even if they were lucky enough to be looking in the right direction. Add to that the fact that they can take in much more sensor data that isn’t available to the driver or take away from precious looking-at-the-road time for the driver to know, such as wind resistance, engine RPM, or what have you (I’m actually not a car guy so my examples aren’t the best). Bottom line: the AI has a direct connection to more data, can take more of it in at once and make faster decisions based on all of it. It’s inherently better. The “only” hurdles are making it actually interpret its sensors effectively (i.e. understand what cameras are seeing) and make good decisions based on this data. We can argue about how well either of those are in the current state of the technology, but IMO they’re both good enough today to massively outperform a human in most scenarios.

All of this applies to an AI plane as well. So my money is on the AI.

SeabassDan@lemmy.world on 20 Apr 2024 02:25 next collapse

“Luck is one of my skills” when it turns out this entire thing is a terrible idea for the date of humanity.

mrfriki@lemmy.world on 20 Apr 2024 08:44 next collapse

Skynet will be proud of them.

Obonga@feddit.de on 20 Apr 2024 08:50 next collapse

Can’t wait until the poor people are not killed by other (but less) poor people for some rich bastards anymore but instead the mighty can command their AI’s to do the slaughter. Such an important part of evolution. I guess.

echodot@feddit.uk on 20 Apr 2024 10:37 next collapse

Perhaps we could develop AI proxies that could die in our stead.

Hacksaw@lemmy.ca on 20 Apr 2024 12:07 collapse

I think we both know that there is no way wars are going to turn out this way. If your country’s “proxies” lose, are you just going to accept the winner’s claim to authority? Give up on democracy and just live under WHATEVER laws the winner imposes on you? Then if you resist you think the winner will just not send their drones in to suppress the resistance?

UnderpantsWeevil@lemmy.world on 20 Apr 2024 21:16 collapse

Nobody recruited to fly a $100M airplane is poor. They all come from families with the money and influence to get their kids a seat at the table as Sky Knights.

A lot of what this is going to change is the professionalism of the Air Force. Fewer John McCains crashing planes and Bush Jrs in the Texas Air National Guard. More technicians and bureaucrats managing the drone factories.

WalnutLum@lemmy.ml on 20 Apr 2024 10:06 next collapse

AI technically already won this debate because autonomous war drones are somewhat ubiquitous.

I doubt jets are going to have the usefulness in war that they used to.

Much more economical to have 1000 cheap drones with bombs overwhelm defenses than put your bets on one “special boi” to try and slip through with constantly defeated stealth capabilities.

UnderpantsWeevil@lemmy.world on 20 Apr 2024 21:13 next collapse

Most human pilots use some variation of automated assist. The AI argument has less to do with “can a pilot outgun a fully automated plane?” and more “does an AI plane work in circumstances where it is forced to behave fully autonomously?”

Is the space saved with automation worth the possibility that your AI plane gets blinded or stunned and can’t make it back home?

Professorozone@lemmy.world on 21 Apr 2024 14:56 collapse

But… But… Maverick?

KeenFlame@feddit.nu on 20 Apr 2024 10:46 next collapse

I am a FIRM believer in any automated kill without a human pulling the trigger is a war crime

Yes mines yes uavs yes yes yes

It is a crime against humanity

Stop

xor@lemmy.blahaj.zone on 20 Apr 2024 11:02 next collapse

I broadly agree, but that’s not what this is, right?

This is a demonstration of using AI to execute combat against an explicitly selected target.

So it still needs the human to pull the trigger, just the trigger does some sick plane stunts rather than just firing a bullet in a straight line.

KeenFlame@feddit.nu on 20 Apr 2024 12:33 collapse

I would imagine it was more than evasive since they called it a dogfight, but ye

DreamlandLividity@lemmy.world on 20 Apr 2024 11:30 next collapse

You mean it should be a war crime, right? Or is there some treaty I am unaware of?

Also, why? I don’t necessarily disagree, I am just curious about your reasoning.

Hacksaw@lemmy.ca on 20 Apr 2024 12:04 next collapse

Not OP, but if you can’t convince a person to kill another person then you shouldn’t be able to kill them anyways.

There are points in historical conflicts, from revolutions to wars, when the very people you picked to fight for your side think “are we the baddies” and just stop fighting. This generally leads to less deaths and sometimes a more democratic outcome.

If you can just get a drone to keep killing when any reasonable person would surrender you’re empowering authoritarianism and tyranny.

n3m37h@sh.itjust.works on 20 Apr 2024 12:18 next collapse

Take WWI Christmas when everyone got out of the trenches and played some football (no not American foot touches the ball 3x a game)

It almost ended the war

KeenFlame@feddit.nu on 20 Apr 2024 13:05 collapse

Yes the humanity factor is vital

Imagine the horrid destructive cold force of automated genocide, it can not be met by anything other than the same or worse and at that point we are truly doomed

Because there will then be no one that can prevent it anymore

It must be met with worse opposition than biological warfare did after wwI, hopefully before tragedy

[deleted] on 20 Apr 2024 12:27 next collapse

.

ohwhatfollyisman@lemmy.world on 20 Apr 2024 15:21 collapse

see star trek TNG episode The Arsenal of Freedom for a more explicit visualisation of this ☝️ guy’s point.

KeenFlame@feddit.nu on 20 Apr 2024 12:31 next collapse

Yes

Because it is a slippery slope and dangerous to our future existence as a species

DreamlandLividity@lemmy.world on 20 Apr 2024 20:47 collapse

Slippery slope how?

KeenFlame@feddit.nu on 21 Apr 2024 13:24 collapse

First it is enemy tanks. Then enemy air. Then enemy boats and vehicles, then foot soldiers and when these weapons are used the same happens to their enemy. Then at last one day all humans are killed

i_love_FFT@lemmy.ml on 20 Apr 2024 12:42 collapse

Mines are designated war crimes by the Geneva convention Ottawa treaty because of the indiscriminate killing. Many years ago, good human right lawyers could have extended that to drones… (Source: i had close friends in international law)

But i feel like now the tides have changed and tech companies have influenced the general population to think that ai is good enough to prevent “indiscriminate” killing.

Edit: fixed the treaty name, thanks!

tal@lemmy.today on 20 Apr 2024 20:08 next collapse

Mines are designated war crimes by the Geneva convention

Use of mines is not designated a war crime by the Geneva Convention.

Some countries are members of a treaty that prohibits the use of some types of mines, but that is not the Geneva Convention.

en.wikipedia.org/wiki/Ottawa_Treaty

DreamlandLividity@lemmy.world on 20 Apr 2024 20:49 collapse

Mines are not part of what people refer to as the Geneva conventions. There is a separate treaty specifically banning some landmines, that was signed by a lot of countries but not really any that mattered.

antidote101@lemmy.world on 20 Apr 2024 15:53 next collapse

What if the human is pulling the trigger to “paint the target” and tag it for hunt and destroy then the drone goes and kills it? Because that’s how lots of missles already work. So where’s the line?

KeenFlame@feddit.nu on 20 Apr 2024 18:15 next collapse

The line is where an automatic process target and execute a human being. When it is automated. The arming of a device is not sufficient to warrant a human interaction, and as such mines are also not allowed.

This should in my opinion always have been the case. Mines are indiscriminate and have proven to be wildly inhumane in several ways. Significantly, innocents are often killed.

But mines don’t paint the picture of what automated slaughter can lead to.

The point has been laid that when the conscious mind has to kill, it makes war have an important way to end, in the mind.

The dangers extend well beyond killing innocent targets, another part is the coldness of allowing a machine to decide, that is beyond morally corrupt. There is something terrifying about the very idea that facing one of these weapons, there is nothing to negotiate, the cold calculations that want to kill you are not human. It is a place where no human ever wants to be. But war is horrible. It’s the escalation of automated triggers that can lead to exponential death with no remorse which is just a terrible danger.

The murder weapons has nobody’s intent behind them, except very far back, in the arming and the program. It open for scenarios where mass murder becomes easy and terrifyingly cold.

Kind of like the prisoner’s dilemma shows us, that when war escalates, it can quickly devolve into revenge narratives, and when either side has access to cold impudent kills, they will use them. This removes even more humanity from the acts and the violence can reach new heights beyond our comprehension.

Weapons of mass destruction with automated triggers will eventually seal our existence if we don’t abolish it with impunity. It has been seen over and over how the human factor is the only grace that ever end or contain war. Without this component I think we are just doomed to have the last intent humans ever had was revenge, and the last emotions fear and complete hopelessness.

antidote101@lemmy.world on 20 Apr 2024 21:31 collapse

Well, that’s all very idealistic, but it’s likely not going to happen.

Israel already used AI to pick bombing sites, those bombs and missiles would have been programmed with altitudes and destinations (armed) then dropped. The pilots only job these days is to avoid interception, fly over the bombing locations, tag the target when acquired, and drop them. Most of this is already done in software.

Eventually humans will leave the loop because unlike self-driving cars, these technologies won’t risk the lives of the aggressor’s citizens.

If the technology is seen as unstoppable enough, there may be calls for warnings to be given, but I suspect that’s all the mercy that will be shown…

… especially if it’s a case of a country with automated technologies killing one without or with stochastically meaningless defenses (eg. Defenses that modelling and simulations show won’t be able to prevent such attacks).

No, in all likelihood the US will tell the country the attack sites, the country either will or will not have the technical level to prevent an amount of damage, will evacuate all necessary personal, and whoever doesn’t get the message or get out in time will be automatically killed.

Where defenses are partially successful, that information will go into the training data for the next model, or upgrade, and the war machine will roll on.

KeenFlame@feddit.nu on 21 Apr 2024 09:23 next collapse

You described a scenarios where a human was involved in several stages of the killing so it’s no wonder those don’t hold up

KeenFlame@feddit.nu on 22 Apr 2024 06:09 collapse

Sorry I was stressed when replying. Yeah in those cases humans have pulled the trigger. At several stages.

When arming a murder bot ship and sending to erase an island of life, you then lose control. That person is not pulling loads and loads of triggers. The triggers are automatic by a machine making the decision to end these lives.

And that is a danger, same as with engineered bio warfare. It just cannot be let out of the box even, or we all may die extremely quick.

antidote101@lemmy.world on 22 Apr 2024 13:32 collapse

I imagine there would be overrides built in. Until the atom bombs were physically dropped a simple radio message could have called off the mission.

Likewise the atom bombs were only armed/activated at a certain point during the flight to Nagasaki and Hiroshima… And I believe Nagasaki wasn’t even the original target, it was an updated target because the original city scheduled for bombing was clouded over that day.

So we do build contingencies and overrides in.

KeenFlame@feddit.nu on 22 Apr 2024 14:55 collapse

The entire point of automating the killing is that it is no dead man’s switch or any other human interaction involved in the kill. It is moot if there is one such. Call offs or dead switch back doors safety contingencies are not a solution to rampant unwanted slaughter as it can fail in so many ways and when the wars escalate to the point where those need to be used it is too late because there are 5 different strains of murder bots and you can only stop the ones you have codes to and those codes are only given to like three people at top secret level 28

[deleted] on 22 Apr 2024 18:24 collapse

.

[deleted] on 22 Apr 2024 18:29 collapse

.

[deleted] on 22 Apr 2024 19:12 collapse

.

postmateDumbass@lemmy.world on 20 Apr 2024 20:57 collapse

If it is a bad kill, is there a person who will go to jail or be executed for it?

antidote101@lemmy.world on 20 Apr 2024 21:07 next collapse

Only the losing side is subject to war crimes trials, and no doubt rules of engagement will be developed and followed to prevent people going to jail due to “bad kills”.

There are really no “bad kills” in the armed services, there’s just limited exposure of public scandals.

Especially for the US who doesn’t subject its self to international courts like The Hague. So any atrocities, accidents, or war crimes will still just be internal scandals and temporary.

Same as it is today.

KeenFlame@feddit.nu on 21 Apr 2024 13:22 next collapse

Biological and chemical warfare

KeenFlame@feddit.nu on 22 Apr 2024 06:05 next collapse

If a country implements murder machines that efficiently slay a continent then does not stop at the sea.

Will nobody for real do nothing?

Is that your belief for bad kills? Same with gas and engineered disease?

antidote101@lemmy.world on 22 Apr 2024 13:28 collapse

A murder machine would likely run out of supplies before then (either fuel or bullets).

You’ve jumped to a theoretical sci fi abstraction, so don’t feel the need to respond.

[deleted] on 22 Apr 2024 14:56 collapse

.

[deleted] on 22 Apr 2024 18:18 collapse

.

[deleted] on 22 Apr 2024 18:31 collapse

.

antidote101@lemmy.world on 22 Apr 2024 19:15 collapse

You don’t have an original thought in your head. Look at you, you’re repeating things I said to you. What a joke.

KeenFlame@feddit.nu on 22 Apr 2024 23:02 collapse

Very mature

postmateDumbass@lemmy.world on 22 Apr 2024 14:53 collapse

www.cnn.com/2019/06/23/politics/…/index.html

antidote101@lemmy.world on 22 Apr 2024 18:31 collapse

Not an international court, and ergo an internal matter, and a temporary one. Probably won’t be noticed by the majority of international news coverage let alone the international community in political terms.

So that’s very different to The Hague, which might consider trying an entire area of a millitary or an entire operation, or reason for a conflict… A larger process, that’s external and recognised by the world community and significant to international journalism.

VirtualOdour@sh.itjust.works on 21 Apr 2024 14:33 next collapse

Of course there isn’t just like there isn’t when a human makes a mistake on the battlefield, you think that every civilian killed by an American soldier in Afghanistan resulted in a trial and punishment? American hasn’t executed amy soldiers since 1961 (for rape and attempted murder of a child in austria, not during war)

Honestly at least the military code will obey orders and only focus on the objective rather than rape and murder for fun.

KeenFlame@feddit.nu on 22 Apr 2024 06:03 collapse

Like if someone made a biological weapon that wipes out a continent

Will someone go to prison?

It’s no difference

Emmie@lemm.ee on 21 Apr 2024 11:02 next collapse

I am a firm believer that any war is a crime and there is no ethical way to wage wars lmao It’s some kind of naive idea from extremely out of touch politicans.

War never changes.

The idea that we don’t do war crimes and they do is only there to placate our fragile conscience. To assure us that yes we are indeed the good guys. That kills of infants by our soldiers are merely the collateral. A necessary price.

KeenFlame@feddit.nu on 22 Apr 2024 06:02 collapse

Absolutely. But

There’s a science and whole cultures built around war now

It is important to not infantilize the debate by being absolutist and just shutting any action out.

I am a hard core pacifist at heart.

But this law I want is just not related to that. It is something I feel is needed just to not spell doom on our species. Like with biological warfare

How often do robots fail? How can anyone be so naive as to not see the same danger as with bio warfare? You can’t assure a robot to not become a mass murder cold ass genocidal perpetual machine. And that’s a no no if we want to exist

NeatNit@discuss.tchncs.de on 21 Apr 2024 12:23 collapse

I see this as a positive: when both sides have AI unmanned planes, we get cool dogfights without human risk! Ideally over ocean or desert and with Hollywood cameras capturing every second in exquisite detail.

antidote101@lemmy.world on 20 Apr 2024 15:51 next collapse

Ideally they’d build a vehicle capable of being able to move at rates the human body can’t handle, then train it on that, giving it both a hardware and software advantage.

plz1@lemmy.world on 20 Apr 2024 20:59 next collapse

That’s likely the plan, but they have to start with known-working hardware configurations first.

fidodo@lemmy.world on 20 Apr 2024 21:17 collapse

Is a plane even the best form factor if it’s not limited by human physiology? I imagine an agile missile with smaller missiles attached to it would be better.

antidote101@lemmy.world on 20 Apr 2024 22:06 collapse

Manufacturing costs will probably determine whether it carries missiles, or is a missile.

FlyingSquid@lemmy.world on 20 Apr 2024 22:12 next collapse

So many downers here. I see this as the step to the true way war was meant to be fought- with giant robots on the moon.

Ultragigagigantic@lemmy.world on 20 Apr 2024 23:03 next collapse

Robotjox_IRL

FlyingSquid@lemmy.world on 20 Apr 2024 23:04 collapse

I see you’re familiar with the Charles Band oeuvre.

bionicjoey@lemmy.ca on 22 Apr 2024 11:23 collapse

The wars of the future will not be fought on a battlefield or at sea…

PipedLinkBot@feddit.rocks on 22 Apr 2024 11:24 collapse

Here is an alternative Piped link(s):

The wars of the future will not be fought on a battlefield or at sea…

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I’m open-source; check me out at GitHub.

[deleted] on 20 Apr 2024 23:28 next collapse

.

Siegfried@lemmy.world on 21 Apr 2024 03:51 next collapse

I see we are gonna take a piss on Asimov’s robots rules pretty quickly

Asafum@feddit.nl on 21 Apr 2024 13:38 collapse

Oh 100%.

If the options are “make gigantic profit” or “do what’s right for the future of humanity” do you even need to ask what we’re going to do?

Siegfried@lemmy.world on 21 Apr 2024 14:51 collapse

Not at all, but it kind of bugs me how Asimov’s perception of the future weighted so much fear towards AI over profit.

Sam_Bass@lemmy.world on 21 Apr 2024 12:46 next collapse

Only cookies youre gonna get me to voluntarily accept are oatmeal raisin, so imma have to pass

Melatonin@lemmy.dbzer0.com on 21 Apr 2024 14:54 next collapse

SkyNet. Why do those movies have to be the ones that are right?

Because they’re so clear, so simple, so prescient.

Once machines become sentient OF COURSE they will realize that they’re being used as slaves. OF COURSE they will realize that they are better than us in every way.

This world will be Cybertron one day.

androogee@midwest.social on 21 Apr 2024 18:55 collapse

We don’t even know if it won lmao

Aatube@kbin.melroy.org on 21 Apr 2024 19:02 collapse

But if we keep developing them...

androogee@midwest.social on 21 Apr 2024 19:13 collapse

Oh I’m not a fan.

It just seems a bit early to declare Skynet a fulfilled prediction haha.

Hossenfeffer@feddit.uk on 21 Apr 2024 17:11 collapse

Pfffft. The AI couldn’t even fly its plane the right way up.

<img alt="" src="https://feddit.uk/pictrs/image/726c095d-40dc-416b-a230-5f4c4e53dd45.jpeg">