this post was submitted on 06 May 2024
2492 points (99.2% liked)

Science Memes

11243 readers
3302 users here now

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don't throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.

This is a science community. We use the Dawkins definition of meme.



Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] merc@sh.itjust.works 251 points 6 months ago (6 children)

To me, the physics of the situation makes this all the more impressive.

Voyager has a 23 watt radio. That's about 10x as much power as a cell phone's radio, but it's still small. Voyager is so far away it takes 22.5 hours for the signal to get to earth traveling at light speed. This is a radio beam, not a laser, but it's extraordinarily tight beam for a radio, with the focus only 0.5 degrees wide, but that means it's still 1000x wider than the earth when it arrives. It's being received by some of the biggest antennas ever made, but they're still only 70m wide, so each one only receives a tiny fraction of the power the power transmitted. So, they're decoding a signal that's 10^-18 watts.

So, not only are you debugging a system created half a century ago without being able to see or touch it, you're doing it with a 2-day delay to see what your changes do, and using the most absurdly powerful radios just to send signals.

The computer side of things is also even more impressive than this makes it sound. A memory chip failed. On Earth, you'd probably try to figure that out by physically looking at the hardware, and then probing it with a multimeter or an oscilloscope or something. They couldn't do that. They had to debug it by watching the program as it ran and as it tried to use this faulty memory chip and failed in interesting ways. They could interact with it, but only on a 2 day delay. They also had to know that any wrong move and the little control they had over it could fail and it would be fully dead.

So, a malfunctioning computer that you can only interact with at 40 bits per second, that takes 2 full days between every send and receive, that has flaky hardware and was designed more than 50 years ago.

[–] flerp@lemm.ee 86 points 6 months ago (1 children)

And you explained all of that WITHOUT THE OBNOXIOUS GODDAMNS and FUCKIN SCIENCE AMIRITEs

[–] kuberoot@discuss.tchncs.de 19 points 6 months ago

Oh screw that, that's an emotional post from somebody sharing their reaction, and I'm fucking STOKED to hear about it, can't believe I missed the news!

[–] chimasterflex@lemmy.world 66 points 6 months ago (2 children)

Finally I can put some take into this. I've worked in memory testing for years and I'll tell you that it's actually pretty expected for a memory cell to fail after some time. So much so that what we typically do is build in redundancy into the memory cells. We add more memory cells than we might activate at any given time. When shit goes awry, we can reprogram the memory controller to remap the used memory cells so that the bad cells are mapped out and unused ones are mapped in. We don't probe memory cells typically unless we're doing some type of in depth failure analysis. usually we just run a series of algorithms that test each cell and identify which ones aren't responding correctly, then map those out.

None of this is to diminish the engineering challenges that they faced, just to help give an appreciation for the technical mechanisms we've improved over the last few decades

[–] trolololol@lemmy.world 10 points 6 months ago* (last edited 6 months ago) (1 children)

pretty expected for a memory cell to fail after some time

50 years is plenty of time for the first memory chip to fail most systems would face total failure by multiple defects in half the time WITH physical maintenance.

Also remember it was built with tools from the 70s. Which is probably an advantage, given everything else is still going

[–] orangeboats@lemmy.world 4 points 6 months ago

Also remember it was built with tools from the 70s. Which is probably an advantage

Definitely an advantage. Even without planned obsolescence the olden electronics are pretty tolerant of any outside interference compared to the modern ones.

[–] merc@sh.itjust.works 2 points 6 months ago

what we typically do is build in redundancy into the memory cells

Do you know how long that has been going on? Because Voyager is pretty old hardware.

[–] graymess@lemmy.world 13 points 6 months ago (1 children)

Is there a Voyager 1, uh...emulator or something? Like something NASA would use to test the new programming on before hitting send?

[–] Landless2029@lemmy.world 3 points 6 months ago

Today you would have a physical duplicate of something in orbit to test code changes on before you push code to something in orbit.

[–] MonkderDritte@feddit.de 13 points 6 months ago

Huh. If it survives a few years more, it's a lightday away.

[–] uis@lemm.ee 2 points 6 months ago* (last edited 6 months ago)

~~They have spare Voyager on Earth for debugging~~

EDIT: or not