• 0 Posts
  • 89 Comments
Joined 1 year ago
cake
Cake day: June 9th, 2023

help-circle

  • aard@kyu.detoProgrammer Humor@programming.devOld timers know
    link
    fedilink
    arrow-up
    3
    arrow-down
    2
    ·
    10 days ago

    Shitty companies did it like that back then - and shitty companies still don’t properly utilize what easy tools they have available for controlled deployment nowayads. So nothing really changed, just that the amount of people (and with that, amount of morons) skyrocketed.

    I had automated builds out of CVS with deployment to staging, and option to deploy to production after tests over 15 years ago.








  • One fascinating example is one owner that replaced the DC barrel jack with a USB-C port, so they could utilize USB-PD for external power.

    Oddly enough that’s also an example for bad design in that notebook: The barrel jack is soldered in. With a module that is plugged into the board that’d be significantly easier to replace - and also provide strain relief for power jack abuse. All my old thinkpads were trivial to move to USB-C PD because they use a separate power jack with attached cable.

    The transparent bottom also isn’t very functional - it is pretty annoying to remove and put back, due to the large amount of screws required. For a notebook designed for tinkering I’d have wanted some kind of quick release for that. Also annoying is the lack of USB ports on the board - there’s enough space to integrate a USB hub, but just doing that on the board and providing extra ports would’ve been way more sensible.

    The CPU module also is a bit of a mixed bag - it pretty much is designed for the first module they developed, and later modules don’t have full support for the existing ports. I was expecting that, though - many projects trying to offer that kind of modular upgrade path run into that sooner or later, and for that kind of small project with all its teething problems ‘sooner’ was to be expected. It still is very interesting for some prototyping needs - but that’s mostly companies or very dedicated hackers, not the average linux user.


  • Admittedly I’m just toying around for entertainment purposes - but I didn’t really have any problems of getting anything I wanted to try out with rocm support. Bigger annoyance was different projects targetting specific distributions or specific software versions (mostly ancient python), but as I’m doing everything in containers anyway that also was manageable.


  • For AI and compute… They’re far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins

    I got a W6800 recently. I know a nvidia model of the same generation would be faster for AI - but that thing is fast enough to run stable diffusion variants with high resolution pictures locally without getting too annoyed.









  • Apple has low memory behaviour way better optimized than Windows, so running at 8GB will not be as painful as it is with windows - but in the background the OS will constantly shuffle stuff around to avoid running out of memory, which costs performance.

    16GB is the bare minimum for computer nowadays - and that applies to macs as well. I’m currently using a 16GB air m1 for some things, and I also regularly run into performance issues due to memory limits without doing heavy stuff.