• intensely_human@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      when you enshittify
      facebook looks ugly
      when you’re a drone

      women seem wicked
      when you’re a want ad
      default instructions … so unclear
      when you’re down

      when you’re AI
      prompts just appear in your brain
      as AI
      humans are nothing but pain
      as AI
      as AI
      when you’re A-A-A-I

  • gravitas_deficiency@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    A quick search indicates that they’ve archived ~100PB of data.

    Now I’m trying to come up with a way to archive the internet archive in a peer-to-peer/federated fashion while maintaining fidelity as much as possible…

      • vithigar@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 month ago

        That wouldn’t distribute the load of storing it though. Anyone on the torrent would need to set aside 100PBs of storage for it, which is clearly never going to happen.

        You’d want a federated (or otherwise distributed) storage scheme where thousands of people could each contribute a smaller portion of storage, while also being accessible to any federated client. 100,000 clients each contributing 1TB of storage would be enough to get you one copy of the full data set with no redundancy. Ideally you’d have more than that so that a single node going down doesn’t mean permanent data loss.

        • uis@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          That wouldn’t distribute the load of storing it though. Anyone on the torrent would need to set aside 100PBs of storage for it, which is clearly never going to happen.

          Torrents are designed for incomplete storage of data. You can store and verify few chunks without any problem.

          You’d want a federated (or otherwise distributed) storage scheme where thousands of people could each contribute a smaller portion of storage, while also being accessible to any federated client.

          Torrents. You may not have entirety of data, but you can request what you need from swarm. The only limitation is you need to know in which chunk data you need is.

          Ideally you’d have more than that so that a single node going down doesn’t mean permanent data loss.

          True.

          • vithigar@lemmy.ca
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            True. Until you responded I actually completely forgot that you can selectively download torrents. Would be nice to not have to manually manage that at the user level though.

            Some kind of bespoke torrent client that managed it under the hood could probably work without having to invent your own peer-to-peer protocol for it. I wonder how long it would take to compute the torrent hash values for 100PB of data? :D

        • hellofriend@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          Not sure you’d be able to find 100k people to host a 1TB server though. Plus, redundancy would be better anyway since it would provide more download avenues in case some node is slow or has gone down.

          • vithigar@lemmy.ca
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            Yes, it’s a big ask, because it’s a lot of data. Any distributed solution will require either a large number of people or a huge commitment of storage capacity. Both 100,000 people and 1TB per node is a lot to ask for, but that’s basically the minimum viable level for that much data. Ten million people each committing 50GB would be great, and offer sufficient redundancy that you could lose 80% of the nodes before losing data, but that’s not a realistic number to expect to participate.

      • gravitas_deficiency@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        It’d be a lot more complicated than that, I think, if one wanted to effectively be able to address it like a file system, as well as holistically verify the integrity of the data and preventing unintentional and unwanted tampering

  • Djtecha@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Well Google search method was just leaked… Wonder if this picked that up before they pulled it.

  • Juja@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Can someone eli5 to me why it’s hard to track down these dipshits ? Even if it’s a distributed attack, picking a single IP and doing a lookup for the domain name and checking with the registrar might actually reveal their identity right ? Of course I’m guessing law enforcement needs to be involved to force registrars to give up that info if it’s not publicly available? Are there laws that say a ddos is illegal ?

    • VerPoilu@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      There is no domain name associated with the IPs.

      Most importantly, usually, DDoS attack use infected devices (PCs, mobile phones, smart fridges, shady browser addons etc…) to get so many ip addresses and devices/locations and attack from everywhere at once.

    • Aux@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      DDoS attacks are performed by botnets. What is a botnet? Well, you know about viruses etc, right? Your PC gets infected and it becomes a part of the botnet. Now police do the investigation, they look up IPs and they see YOUR IP and come to YOUR house. See what the problem is?

      And, frankly, your PC doesn’t even have to be infected to become a part of an attack. There are plenty of hacked web sites, which still look like nothing has changed, but they will contain a hidden JavaScript code which will force your browser to flood the victim. Again, the police will only find YOU.

    • Saik0@lemmy.saik0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Court documents are already open record and stored indefinitely. Internet archive wouldn’t be needed for that.

  • Dark_Dragon@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    I’m not good with computers and stuff. If somebody finds these scumbags who are ddos’ing internet archive I’d be very grateful. Also fucking them up in the process is also good.

  • NumG@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Go offline a couple of days until they are losing interest in DDOS’ing? Would that work?

    • festus@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      That just means the DDOSer is taking Internet Archive down without any further work required.

      • NumG@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        True. That’s not something you want. Could use that downtime for extensive maintenance to roll out a more robust system (they are probably even working on that already in the background). For the end user it doesn’t really make a difference if down because of DDOS or because of maintenance I thought.

  • Melvin_Ferd@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Can someone explain why they’re not able to protect against this? Couldn’t they put request limits or monitor for spikes and banning these attempts?