Full scan of 1 cubic millimeter of brain tissue took 1.4 petabytes of data, equivalent to 14,000 4K movies — Google's AI experts assist researchers (www.tomshardware.com)
from KarnaSubarna@lemmy.ml to technology@lemmy.ml on 11 May 2024 18:37
https://lemmy.ml/post/15523078

#technology

threaded - newest

SouthFresh@lemmy.ml on 11 May 2024 19:01 next collapse

“Google to shutter human brains”

Why anyone teams up with a company with its greatest achievement being a high score on the “I wish they hadn’t shut that down” list, is beyond my understanding.

GolfNovemberUniform@lemmy.ml on 11 May 2024 20:05 next collapse

Because almost nobody else has enough money for such research and governments won’t pay for it because it’s not very useful

fruitycoder@sh.itjust.works on 12 May 2024 02:31 collapse

I mean kubernetes, android, tensorflow, and the only OpenSource PDK for silicon that I know of.

They have a lot of bed rock contributions in the tech space.

chirospasm@lemmy.ml on 11 May 2024 19:12 next collapse

“We did the back-of-napkin math on what ramping up this experiment to the entire brain would cost, and the scale is impossibly large — 1.6 zettabytes of storage costing $50 billion and spanning 140 acres, making it the largest data center on the planet.”

Look at what they need to mimic just a fraction of our power.

henfredemars@infosec.pub on 11 May 2024 21:51 next collapse

It’s not even complete. You might have the physical brain tissue, but that tissue is stateful. The tissue contains potentials and electrical charges that must be included in a complete model.

zagaberoo@beehaw.org on 11 May 2024 23:50 next collapse

And the whole human body, brain and all, can run on ~100 watts. Truly astounding.

[deleted] on 12 May 2024 00:31 next collapse

.

whoisearth@lemmy.ca on 12 May 2024 01:46 collapse

Meh. No different than how you can make a programming language and then have it compile itself. It’s weird.

[deleted] on 12 May 2024 05:44 collapse

.

BastingChemina@slrpnk.net on 13 May 2024 06:38 collapse

And we get these 100W from transforming food that we find on/in the ground.

UnfortunateShort@lemmy.world on 11 May 2024 23:35 next collapse

Heh, we the best. Now excuse me while I use my amazing brainpower to watch questionable anime and get more depressed 🫥

whoisearth@lemmy.ca on 12 May 2024 01:43 collapse

Every day we stray further from God’s light lol

Zerfallen@lemmy.world on 12 May 2024 10:19 collapse

naruto-running from his light

Midnitte@beehaw.org on 12 May 2024 00:47 next collapse

Sort of makes you realize how daunting self driving cars are.

ThePantser@lemmy.world on 12 May 2024 01:36 next collapse

Shouldn’t be long and we will have that much in our phones.

SirEDCaLot@lemmy.today on 12 May 2024 02:41 next collapse

In fairness, the scan required such astronomical resources because of how they were scanning it. They took the cubic millimeter chunk and cut it into 5,000 super thin flat slices and then did extremely high detail scans of each slice. That’s why they needed AI, to try and piece those flat layers back together into some sort of 3D structure.

Once they have the 3D structure, the scans are useless and can be deleted.

In time it should be possible to scan the tissue and get the 3D structure without such extreme data use.

redcalcium@lemmy.institute on 12 May 2024 09:58 collapse

Imagine donating your body to science and the scientists slice your brain and scan them, then decades later you suddenly wake up in a virtual space because the scientists are finally able to emulate a copy of your brain in a supercomputer.

SirEDCaLot@lemmy.today on 12 May 2024 12:08 next collapse

Sounds good to me. You should also look into cryonics. Basically you sign up with a company and donate your body to them, when you die they pump you full of antifreeze and then vitrify you in liquid nitrogen. Right now there’s no way to recover from it, the antifreeze is toxic and we don’t yet know how to undo the cell damage from freezing. But the idea is someday in the future we will figure those things out, and then hopefully be able to thaw the frozen dead person, fix the damage caused by the freezing process, fix whatever problem killed them in the first place, and reanimate them.

For a lower fee, they will cut off your head and just freeze that. Idea being that someday in the future they will be able to transplant your brain into an artificially created body.

TheRealKuni@lemmy.world on 13 May 2024 13:19 collapse

But they really only unfreeze people who knew Richard Dawkins and Mrs. Garrison. Then laugh at you when you want to play Nintendo Wii.

USSEthernet@startrek.website on 12 May 2024 19:47 collapse

We are Bob

Korkki@lemmy.world on 12 May 2024 04:12 next collapse

it’s not like human brain memory or consciousness is that information dense. They just did that high of a definition of a scan.

Aopen@discuss.tchncs.de on 12 May 2024 17:48 collapse

CERN datacenter has 1600 times less capacity

home.cern/news/news/…/exabyte-disk-storage-cern

Although global storage capacity will be 125 times higher by 2025 than whole scan would occupy

cybersecurityventures.com/the-world-will-store-20…

utopiah@lemmy.ml on 13 May 2024 13:09 collapse

I’d be curious about the access speed comparison, because I’d assume for the brain it’s be RAM equivalent, not SDD

Vivendi@lemmy.zip on 13 May 2024 13:17 next collapse

The brain is a tightly coupled biological computer , it’s access speed is practically instantaneous

Also data/processing in the brain is some mighty uncovered field of science

TheRealKuni@lemmy.world on 13 May 2024 13:17 collapse

Just gotta lower the clock speed enough for us not to notice. As long as we don’t interact with the outside world, just other stored human brains, it can be slow as molasses and we won’t notice.

riplin@lemm.ee on 11 May 2024 19:27 next collapse

That’s capturing everything. Ultimately you need only a tiny fraction of that data to emulate the human brain.

Numenta is working on a brain model to create functional sections of the brain. Their approach is different though. They are trying to understand the components and how they work together and not just aggregating vast amounts of data.

kakes@sh.itjust.works on 11 May 2024 20:03 next collapse

Of course, not to say the data isn’t also important though. It’s very possible that we’re missing something crucial regarding how the brain functions, despite everything we know so far. The more data we have, the better we can build/test these more streamlined models.

andrew_bidlaw@sh.itjust.works on 11 May 2024 21:16 collapse

These models would likely be tested against these real datasets, so they help each other.

GolfNovemberUniform@lemmy.ml on 11 May 2024 20:04 next collapse

I don’t think any simplified model can work EXACTLY like the real thing. Ask rocket scientists

riplin@lemm.ee on 11 May 2024 21:42 next collapse

Given the prevalence of intelligence in nature using vastly different neurons I’m not sure if you even need to have an exact emulation of the real thing to achieve the same result.

FaceDeer@fedia.io on 12 May 2024 01:19 collapse

Fortunately it doesn't have to be exactly like the real thing to be useful. Just ask machine learning scientists.

eleitl@lemmy.ml on 11 May 2024 20:27 next collapse

No, that captures just the neuroanatomy. Not the properties like density of ion channels, type, value of the synapse and all the things we don’t know yet.

can@sh.itjust.works on 11 May 2024 21:19 next collapse

Ultimately you need only a tiny fraction of that data to emulate the human brain.

Point for simulation theory.

henfredemars@infosec.pub on 11 May 2024 21:52 next collapse

No it does not. It captures only the physical structures. There’s also chemical and electrical state that’s missing.

biscuitswalrus@aussie.zone on 12 May 2024 19:25 collapse

Think of this:

You find a computer from 1990. You take a picture (image) of the 1KB memory chip which is on a RAM stick, there are 4 RAM sticks. You are using a DSLR camera. Your image in RAW comes out at 1GB. You project because there’s 8 chips per stick, and 4 sticks it’ll 32GB to image your 4KB of RAM.

You’ve described nothing about the ram. This measurement is meaningless other than telling you how detailed the imaging process is.

umbrella@lemmy.ml on 11 May 2024 23:00 next collapse

i mean they probably use vast amounts of data to learn how it all works.

eguidarelli@lemmy.world on 12 May 2024 00:04 next collapse

Never seen Numenta talked about in the wild! Worked with them on a pattern recognition project in college and it was freaky similar to how toddlers learned about the world around them.

tsonfeir@lemm.ee on 12 May 2024 04:15 next collapse

I want the whole brain, not a republican.

remotelove@lemmy.ca on 12 May 2024 05:37 collapse

Ultimately you need only a tiny fraction of that data to emulate the human brain.

I am curious how that conclusion was formed as we have only recently discovered many new types of functional brain cells.

While I am not saying this is the case, that statement sounds like it was based on the “we only use 10% of our brain” myth, so that is why I am trying to get clarification.

riplin@lemm.ee on 12 May 2024 06:00 next collapse

Oh I’m not basing that on the 10% mumbo jumbo, just that data capture usually over captures. Distilling it down to just the bare functional essence will result in a far smaller data set. Granted, as you noted, there are new neuron types still being discovered, so what to discard is the question.

MajorSauce@sh.itjust.works on 12 May 2024 10:55 next collapse

Not taking a position on this, but I could see a comparison with doing an electron scan of a painting. The scan would take an insane amount of storage while the (albeit ultra high definition) picture would fit on a Blu-ray.

biscuitswalrus@aussie.zone on 12 May 2024 19:20 collapse

They took imaging scans, I just took a picture of a 1MB memory chip and omg my picture is 4GB in RAW. That RAM the chip was on could take dozens of GB!

comador@lemmy.world on 11 May 2024 19:02 next collapse

Storage vendors are rolling their hands in delight while systems administrators, particularly backup admins are cringing at the thought.

fruitycoder@sh.itjust.works on 12 May 2024 02:28 collapse

A SUV full of tape might have the band with needed to restore from backup, but bless the tops that gotta load those tapes

Lojcs@lemm.ee on 11 May 2024 20:28 next collapse

So a 4k movie is 100 GB? 2 hour movie would make it 110 mbps. Insane bitrate even for h.254 imo

onion@feddit.de on 11 May 2024 22:26 next collapse

The movies as shown in cinema are ~600GB

Tempo@lemmy.ml on 11 May 2024 23:08 next collapse

4K Blu Rays encoded in H265 are usually on 100gb discs, so I can see where they’re coming from

cmnybo@discuss.tchncs.de on 12 May 2024 04:27 collapse

4K bluray can be up to 144 mbps, so that’s reasonable.

morbidcactus@lemmy.ca on 12 May 2024 20:24 collapse

I have LOTR directors cut on my server, haven’t bothered reencoding it because I’m not super experienced with keeping hdr 10 going to h265 or equivalent. Return of the king alone is around 130 gigs across two files, jellyfin says its bitrate is about 70 mbps.

Titanic is only about 74 gigs

Lojcs@lemm.ee on 13 May 2024 01:09 collapse

Maybe it’s just me but for 4k60fps h.265 video above 20 mbps looks indistinguishable to me unless I pause on a frame side by side to compare. I’m not sure about h.264 but it can’t be too many times worse

morbidcactus@lemmy.ca on 13 May 2024 01:34 collapse

I compressed some of the 4k rips I did, all my dvd and 1080p blurays, it’s the HDR only that’s stopped me from some of them as I found I lost it with the settings I was using and I put it on the “list of things I’ll come back to later” shelf.

I recall some banding on a few of the dvd rips, probably was a little too aggressive with the settings I used, but they’re still definitely watchable

Evil_Shrubbery@lemm.ee on 12 May 2024 11:46 next collapse

Yes, humans kinda brute-forced intelligence with current assets - made it bigger (with some birthing issues) & more power hungry (with some cooling issues), but it mostly works.

delirious_owl@discuss.online on 12 May 2024 15:35 collapse

How much of that is dead pixels?