• DannyBoy@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    6 months ago

    Seems like it’s cheap to start the bidding at $2500 but the cheapest thing is probably the initial purchase price after moving it, buying the needed cabling, and electricity bills.

    • mikyopii@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      6 months ago

      I bet manpower costs are significant as well. How many people are needed to run this thing? You probably need engineers with an esoteric set of skills to put it back together and manage it which would not be cheap.

      • gregorum@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        It may be running SLED, but just imagine all the specialized, tweaked af code running on top. They didn’t just pop in a LiveCD and click “Install”.

        • Billiam@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          6 months ago

          They didn’t just pop in a LiveCD and click “Install”.

          Obviously not. In 2017, they would have used a live USB thumbdrive instead of a CD.

        • w2tpmf@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          6 months ago

          No, they probably had to pop the live CD into each node individually and click “instal”. Then run a script on each one to join it to the cluster.

          • Almrond@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            6 months ago

            Kind of, you would use a deployment node to manage the individual blades, they are running really specialized software that is basically useless without the management nodes. It wouldn’t be difficult to spin it up (Terascale would have it ready to batch out jobs within a few hours) but you are going to need to engineer your building around it to even get that far. Your foundation needs to support multiple tons of weight, be perfectly level, be able to deliver megawatts of power, remove megawatts of heat (it is water cooled, so you need to have infrastructure and cooling towers to handle that), and you need to be able to get it into the building to begin with. I have worked on this system a few times, just moving it would literally cost upwards of 7 figures. The computer is pretty easy to use, it’s all of the supporting infrastructure that will need a literal team of engineers. I could (and have, kind of) spin the machine up to start crunching data within a day on my own. Fuck moving it, and double fuck re-cabling it. Literal miles of fiber in those racks.

            You do literally pop in an image that is pre-configured in and it deploys to everything at once. That’s probably the easiest part of the whole setup.

          • gregorum@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            6 months ago

            Of course. I was obviously referring to what it takes to operate it after that. Not to mention how complicated setting that whole mess up is.

      • DannyBoy@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        That I get, but I’m sure the reserve isn’t that high if the starting bid is at $2500. It just seems low for the $30,000,000 the computer cost in 2017.

  • nyan@lemmy.cafe
    link
    fedilink
    English
    arrow-up
    0
    ·
    6 months ago

    Let’s just hope that it isn’t bought up by Bitcoin miners . . .

  • BilboBargains@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    6 months ago

    It’s kind of lame that they need to junk the entire apparatus after only a decade. I get that processor technology moves on apace but we already know it does that so why doesn’t a universal architecture exist where nodes can be added at will?

    • trolololol@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      6 months ago

      If you have too many “slow” modes in a super computer you’ll hit a performance ceiling where everything is bottle necked by the speed of things that are not the CPU: memory, disk for swap, and network for sending partial results across nodes for further partial computing.

      Source: I’ve hang up too much around people doing PhD thesis in these kinds of problems.

      • BilboBargains@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        I would imagine it’s very difficult to make a universal architecture but if I have learnt anything about computers it’s that the manufacturers of software and hardware deliberately created opaque and monolithic systems, e.g. phones. They cynically insert barriers to their reuse and redeployment. There’s no profit motive for corporations to make infintitely scalable computers. Short sighted greed is a much more plausible explanation.

      • afraid_of_zombies@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        One of the reasons why I work in industrial controls. A good day is me sneaking in tech that came after the year 2000. Employment for life and I get to branch out to related stuff. Employer is paying me to take ME and chem-e classes now.

        I don’t know why anyone would spend their life chasing the newest fad tech when you can pick a slow moving one, master it, and master the ones around it. Would much rather be the person who knew how the entire system works vs knowing the last 8 programming languages/frameworks only 1 of which is relevant.

        But hey glad there are people who decide on that lifestyle I like having a better cellphone every year.

    • Almrond@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      6 months ago

      It’s more of an operating cost issue. It’s almost decade-old hardware. It was efficient in its day, but compared to new hardware it just costs so much to run you would be better served investing in something with modern efficiency. It won’t be junked, it will be parted out. If you are someone that wants a cheap homelab with infiniband and shitloads of memory you could pick up a blade for a fraction of what it would otherwise cost. I fully expect it to turn into thousands of reasonably powerful servers for the prosumer and nerd markets instead of running as a monolithic cluster.

  • Almrond@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    6 months ago

    Hey, I have worked on this exact machine before, neat to see they are finally decommissioning it. It would be a terrible purchase to actually use these days though, for the cost of moving and deploying it you could rock a few Hopper or Grace clusters that would outperform the cluster for less than half of the operating overhead.

    I fully expect it to get parted out, the actual components would be far more useful on their own as cheap homelab systems, and would be a much better ROI versus using it as is. This thing is water cooled, just the plumbing would be a nightmare to deal with if you aren’t set up for it, and if you are you would be better off going with a modern architecture anyway.

      • Almrond@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        We were running meteorological models mostly, but I did have a colleague that was trying to use it to predict wildlife migratory patterns using topographical mapping. It was batched out on a few projects at any given time while I was there, it was essentially timeshares between a few different research departments.

          • Almrond@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            6 months ago

            Haha, unfortunately no. None of the blades used a windowing system, so we technically wouldn’t have been able to as there is no graphical output (well, the IPMI controllers could have, but that’s kind of cheating). Although, as I’m thinking about it… We probably could have run it over ASCII graphics in a terminal… Man, that was a bit of a wasted opportunity, weather modelling is boring as hell.

  • mox@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    6 months ago

    Power consumption: 1.7 MW

    I hope it stays decommissioned. We’re burning up the planet too fast already.

    • ivanafterall@kbin.social
      link
      fedilink
      arrow-up
      0
      ·
      6 months ago

      Currently around $51k, reserve not yet met. Buyer responsible for transportation and cabling not included, fyi.