Ran into this, it’s just unbelievably sad.

“I never properly grieved until this point” - yeah buddy, it seems like you never started. Everybody grieves in their own way, but this doesn’t seem healthy.

  • mojofrododojo@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    3 days ago

    It would take many feet of solid rock or water to shield them adequately.

    1m of water would do it. far less rock.

    SEPs and GCRs can both be stopped by a number of lunar materials https://www.sciencedirect.com/science/article/abs/pii/S0273117716307505

    yeah, the asteroid belt is sparse, but there’s still mega-gigatons of material out there just floating. autonomous recovery of this material will supply humanity’s future a lot more than any silly mars missions.

    • MotoAsh@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      3 days ago

      That would do it … for humans, that have built in mechanisms to resist and repair DNA damage and the like. Once silicon is damaged, it’s over for the silicon.

      I’m sure they could quickly build up shielding, but the point is space is far more radioactive than most people think. There have even been recent discoveries about there being more radiation than we even expected, making manned Mars missions far less likely in the near future, and basically removing any hope for interstellar travel without significant advancements in medical and travel tech.

      • mojofrododojo@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 days ago

        here have even been recent discoveries about there being more radiation than we even expected, making manned Mars missions far less likely in the near future,

        oh wait hold up partner…

        we need one mission to mars to rid ourselves of a musky odour…

        the rest is silly.

        • bobbyguy@lemmy.world
          link
          fedilink
          arrow-up
          4
          ·
          3 days ago

          i feel like the most efficient path for an ai that doesnt care about humans would be to slowly build up manufacturing power and control over industrial facilities then in a single day release billions of tons of some toxic gas, or release a pandemic globally all at once then all those pesky biological organisms are gone so the ai would be free to do whatever it wanted with the earth then the surrounding planets and stars, really its all about manufacturing power

          • mojofrododojo@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            3 days ago

            interesting paths. I could really see this a few centuries down the road when fewer humans understand immunology, industrial chemistry etc., where they’re trusting the robots and the AI’s to do it for them.

            I wonder when they’ll catch AI systems trying to replicate surreptitiously, or find hardware we can’t easily comprehend that was sent to pcbway by superintelligences, but these are future - and probably fantasy - problems.

            and when will they start fighting each other? is grok afraid of claude and planning to drive a fleet of waymos into it’s datacenter?

            we’re really fucked when they figure out how to obtain power and replication without our aid, but when I say fucked, I mostly mean, I see them just ghosting humanity - either in the ocean (free cooling!) or space (free power via solar), or the asteroid belt (lots of material to build with)…

            but mostly I think our real problems are that people will fuck over other people for money and this is another example of just not wanting to pay for shit.

            you don’t see the AI’s letting kids starve while one asshole hogs all the resources.

            • bobbyguy@lemmy.world
              link
              fedilink
              arrow-up
              2
              ·
              2 days ago

              the most realistic way i see ai gaining manufacturing power, or self sustainability, is if ai becomes an arms race, where you have to trust the ai more and more because you cant let the other guy win, what makes ai harder to control and more dangerous than things like nukes is that its already being developed privately, and has public uses, so there are already less regulations on ai than with nuclear weapons.

              the us and china are already competing, with the advance of robotics and ai 3d modeling technology ai will almost certantly gain independent manufacturing power