• 0 Posts
  • 33 Comments
Joined 1 year ago
cake
Cake day: July 22nd, 2023

help-circle









  • It really depends on what kind of applications you’re talking about. There are still a number of things it can’t run (or well, probably without a lot of meddling around to get there) in the professional space, like CAD. Hopefully this will change over time.

    For a lot of these products there are free alternatives available, but they often don’t cut the mustard and/or aren’t worth retraining for.

    Another thing you should consider before choosing Linux is hardware support. This is often lacking in Linux. For example, your fancy tablet might work fine as a tablet, but if you want to configure anything about it you might need windows depending on the device.

    The good news is, you can try it without worrying about harming your windows install by doing it say on a usb stick or hdd. It’ll only cost you time and effort.


  • Community support is a thing, it’s not the lack of support that’s to blame here - have you ever used Microsoft support? Linux support is much more accessible even.

    A lot of the blame here, is Microsoft’s clever marketing campaign providing windows to educational institutions - with support - for far below cost, in the early days when pc adoption was on the rise.

    Distribution saturation is a barrier to entry and focused support, and it is sometimes more complicated to install and repair. Sometimes it’s easier to repair, because windows is too busy trying to hide its internals from you.

    It’s usually easier to support a remote IT-illiterate person using Linux, by comparison to windows, today.

    e: I guess to be fair, if you factored in community support for windows, your options open up quite a lot. I was more thinking about my own interactions with their support. But enterprise support/problems are not the same as personal ones.



  • I’m going to take the time to illustrate here, how I can see LLMs affecting human speech through existing applications and technologies that are (or could) be made both available and popular enough to achieve this. We’re far enough down the comment chain I can reply to myself now right?

    So, we can all agree that people are increasingly using LLMs in the form of chatgpt and the like, to acquire knowledge/information. The same way as they would use a search engine to follow a link to that knowledge.

    Speech-to-text has been a thing for at least 3 decades (yeah it was pretty hopeless once, but not so much now). So let’s not argue about speech vs text. People already talk to Google and siri and whoever else to this end, llms. Pale have their responses read out via tts.

    I remember being blown away watching a blind sysadmin interacting with a Linux shell via tts at rates I couldn’t even understand the words in 1998. How far we’ve come. I digress, so.

    We’ve all experienced trouble getting the information we’re looking for even with all these tools. Because there’s so much information, and it can be very difficult to find the needle in the haystack. So we constantly have to refine our queries either to be more specific, or exclude relationships to other information.

    This in turn, causes us to think about the words we were using to get the results we want, more frequently because otherwise we spend too much time on recursion.

    In turn, the more we do this, and are trained to do this, the more it will bleed into human communication.

    Now look, there is absolutely a lot of hopium smoking going on here, but damn, this could have everlasting impact on verbal communication. If technology can train people - through inaccurate/incorrect results to think about the communication going out when they speak, we could drastically reduce the amount of miscommunication between people by that alone.

    Imagine:

    get me a chair

    wheels out an office chair from the study

    no I meant a chair for at the kitchen table

    Vs

    get me a chair for at the kitchen table

    You can apply the same thing to human prompted image generation and video generation.

    Now… We don’t need llms to do this, or know this. But we are never going to achieve this without a third party - the “llm”, and whatever it’s plugged into - because the human recipient will usually be more capable of translating these variances, or employ other contexts not as accessible via a single output as speech or text.

    But if machines train us to communicate out better (more accurately, precisely and/or concisely), that is an effect I can’t welcome enough.

    Realistically, the machines will learn to deal with us being dumb, before we adapt.

    e: formatting.


  • This is interesting and thought provoking discussion, ty.

    You’re absolutely right, I was looking for the dead end - plugging LLM into a solution.

    I’m more thinking LLMs used in conjunction with other tech will have these effects on our communicating. LLMs, or whatever replaces them to do that interpretation, are necessary to facilitate that.

    When we come up with something better, to do the same job better, then of course, LLMs will be redundant. If that happens, great.

    We are already seeing a boom in popularity of LLMs outside of professional use. Global ubiquity for anything is never going to happen, unless we can fix communication, which we probably can’t. We certainly can’t alone. It’s very much a chicken an egg problem, that we can only gain from by progressing towards.

    Imagining vocallising using programming languages gave me a chuckle. I have been known to do things like use s/x/y/ to correct in written chats though.

    Programming languages allow us to talk to and listen to machines. LLMs will hopefully allow machines to listen and talk to/between us.




  • Do you actually believe this?

    LLMs are the opposite of a dead end. More like the opening of a pipe. It’s not that they will burn out, it’s just that they’ll reach a point that they’re just one function of a more complete AI perhaps.

    At the very least they tackle a very difficult problem, of communication between human and machine. Their purpose is that. We have to tell machines what to do, when to do it, and how to do it. With such precision that there is no room for error. LLMs are not tools to prove truth, or anything.

    If you ask an LLM a question, and it gives you a response that indicates it has understood your question correctly, and you are able to understand its response that far, then the LLM has done it’s job, regardless of if the answer is correct.

    Validating the facts of the response is another function again, which would employ LLMs as a translation tool.

    It’s not a long leap from there to a language translation tool between humans, where an AI is an interpreter. deepl on roids.