- cross-posted to:
- technews
- cross-posted to:
- technews
Today i was doing the daily ritual of looking at distrowatch. Todays reveiw section was about a termal called warp, it has built in AI for recomendations and correction for commands (like zhs and nushell). You can also as a chatbot for help. I think its a neat conscept however the security is what makes me a bit skittish. They say the dont collect data and you can check it aswell as opt out. But the idea of a terminal being read by an Ai makes me hesitant aswell as a account needed to use warp. What do you guys think?
Sound like LLMs to me. This is not going to stop being a problem. This is the fundamental problem with LLMs - they are text prediction algorithms and have no comprehension of their output.
I’m not sure. Afaik the research is happening. And AI related stuff always happens faster than I can imagine. Ultimately I want the LLMs to hallucinate. They should be able to combine ideas and come up with new and creative answers and be more than just autocomplete. I think what we need is the LLM knowing what it knows and what is made up, and a setscrew. I can see this happening with a higher level of intelligence and/or a clever architecture. I’m not an expert on machine learning myself, however that is what I took from news, companies struggling with their chatbots and everyone wanting their AI assistant to provide factual information. And I don’t see anything ruling that out completely. I mean we humans also sometimes get things wrong or mis-adjust our level of creativity. But I think the concept of facts can be taught to LLMs to some degree, they already seem to grasp it. And concepts have been proposed and things like AI agents that come up with ideas and other agents that check for factuality are in active use. Along with the big tech companies making their AIs cite the sources. In my eyes, progess is being made.
But this is why I currently don’t use LLMs for important and unsupervised stuff, and i try to avoid them when I need correctness. However… I really like to tinker with them, do AI assisted storywriting, or have them come up with 5 creative ideas for a birthday party for my wife. That works well, and with a bit of trickery you can make them output more than the most obvious ideas. And I’m impressed by their ability to code, but as I said it’s still far away from being useful to me. I currently don’t fear for my job. And I additionally struggle with the size of models I can do inference with and their respective intelligence… We’re in the Linux community here, so I think I can be open… I don’t like big tech companies doing my compute and providing me with closed and proprietary services. I don’t use ChatGPT, only open-weight models I can run myself. They aren’t as smart, but I don’t want the future of humankind to be shaped by services and good will of big tech companies.