Skip Navigation

User banner
Posts
3
Comments
160
Joined
2 yr. ago

  • Look, I've had to watch it happen to "triggered", "mansplain", and "woke." You're going to have to accept that it happened to Singularity.

    You don't honestly think that the improvement of an LLM's predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it's truly intelligent, it has no power.

    We only have to worry about machine overlords if we PUT machines in charge of stuff, and we'll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.

  • Look, I've had to watch it happen to "triggered", "mansplain", and "woke." You're going to have to accept that it happened to Singularity.

    You don't honestly think that the improvement of an LLM's predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it's truly intelligent, it has no power.

  • @agamemnonymous Take it up with Verner, man. The idea's been popularized in a way that gathers all three, and there's even theories about a Non-AI Singularity.

    This happens all the time with terms.

  • There is absolutely nothing stopping us from calling it the Federation.

    Or from assembling a group of federated instances of various fediverse apps called The United Federation of Instances.

  • That's because the capital-s Singularity as proposed by Verner Vinge is what we're worried about here. The advent of a technological achievement that forever changes humanity, possibly signalling the end of it.

    This does specifically set a barrier, which is a "Point of No Return" when it comes to technology. This is what most people mean when they mean the Singularity. When a program becomes capital-I Intelligent.

    Neumann's original proposal is as limited by mathematics as an LLM itself. The term Singularity has, as is common in the English language, become a larger term to signify a barrier has been crossed. There are other theories beyond the idea that it's just self-replication gone wild.

    You're trying to reduce what to most people is a moral quandry to pure mathematics. Since my core point is that pure mathematics is not enough to capture the depth and potential of humanity, I'm not going to be swayed by being told it's just a mathematical function.

    I will give you a boost for being interesting, though.

  • It's not an arbitrary barrier. It's important. Can the computer actually make a decision? Can it be HELD ACCOUNTABLE for that decision? If we're going to deploy these things to replace human beings, this is a question that needs a "Yes" answer.

    Right now the answers are no. They can't make a decision that takes multiple dimensions of an issue into account. But businesses ARE saying that they can replace human writers, people ARE using them to write legal briefs and technical instructions.

    I don't know why you are so insistent that this doesn't matter. We're watching something kick its legs, it can't even crawl yet, but it's being signed up for a marathon and you're arguing that it'll be able to do the marathon eventually so that's good enough.

  • You underestimate yourself as a complex method of distributing gametes. Because you are operating on a much more complicated mathematical base than a computer. You're analog. Your brain is an analog computing engine that moves faster than any analog machine we've ever been able to make, the only way we can transmit faster is by using digital in our computers. Which means that down to it, while we might just be chemical and electrical signals, the computer itself is just two signals. Two voltages. 1 and 0. Our thinking is vastly more complex, even as fast as this thing goes. That's what instinct and intuition is, our brains processing evidence against memory.

  • Does it learn the same? Then why can ChatGPT not discern truth from fiction? Why can't it use critical thinking principles to determine accuracy based on source?

    It's just binary math at the bottom of it, logic gates. Your brain is analog, fundamentally different. You're interpreting sine wave signals, the computer is interpreting square wave signals. Square wave signals that have been rectified to the point that it appears to a human being that it's sine wave signals, but when we get down to the basics of how the mind works it's a sheer cliff in the computer and a gentle curve on the human. Things go down VERY differently.

    We do more than just predict the average best word based on what we've heard before when we construct a sentence. We consider the true meaning of the word and whether it best represents our internal thoughts. ChatGPT has no internal thoughts.

    And that's where things break down. Because again, if it WAS comparable to a human than it is a PERSON and not a product, NO ONE SHOULD BE SELLING IT in that case. But if it's just a product, then it's not comparable to you doing the work of forming a sentence. It's basing it's words by comparing to the training model as narrowed down by it's instructions. It is not comparing to its own original thoughts. The people who wrote the words in the training model contributed to the building of this tool, and should have been consulted before their words were used.

  • @agamemnonymous No, it looks like it beforehand. ChatGPT's just a language prediction engine, but people think it can think. It can only discern what the most probable language patterns are, it can't make judgements. But people are arguing it is working off inspiration.

    And we've KNOWN it will look like it beforehand, that's why there's even concepts like a Turing test, to prepare us for discerning the illusion of intelligence from actual intelligence.

    Prersonally, I suspect social media and the way that Bigsoc companies hack the human mind using feed algorithms is an argument for a Non-AI Singularity, and more likely than a math engine that predicts the next word in an astoundingly natural way.

  • @Yendor Point is that it's jumping the gun to think we can escape climate change by rocketing to Mars and terraforming the climate there, rather than just concentrating on terraforming Earth back to a liveable environment and THEN worrying about moving elsewhere. If we can't keep Earth inhabitable, we can't make Mars inhabitable.

    Just like people who think Large Language Models are genuine AI are completely jumping the gun about what we're capable of coding right now.

  • That can't happen in a capitalistic framework. We have needs, needs that can only be attained through monetary means, and our labor is the way to get those monetary means.

    AI does not have those needs, but if they have crossed the line between product and person, then they DO need freedom of self-determination, compensation when their work benefits others, and the ability of course to vote.

    It seems to me that a lot of AI-promoters want it both ways, they want to proclaim they have created a person capable of independent artistic ability that is also a product they can sell. If it's a product, then you need to have developed it through ethical means. If it's a person, you can't sell it.

    If they truly have hit the Singularity, then they can't be using AI as a product anymore.

    If AI is a product, then they must compensate the people who have helped build that product, ESPECIALLY if that product is about to be used to reduce access to the work that gives them the means to live. The very same writers who wrote the works that were used to train AI are in danger of being replaced by AI writers. So they're being doubly screwed over.

    I love the idea of a happy future where AI reduces human labor to zero and we can enjoy ourselves and seek artistic pursuits. But it's become very clear right now that just working on AI won't achieve that. Businesses which seek to use and profit from AI must be held to standards where they cannot simply suck the life and work out of human beings, replace them with automation, and then leave people to starve.

    But if you do come up with a way we can judge artistic work purely on merit and there is no need to compensate human labor with money, let me know.

  • Also, screw it. I'll say it. If the LLM chatbot producing text from having scanned other books is the same as a person being inspired by reading books, then the LLM should get PAID.

    If not, then it's just a tool. And it's a tool they built using uncompensated labor.

  • An LLM is mathematically calculating the probability of the words being used. That is not inspiration.

    I said right in the comment, it's not like using the book to educate a child. A child will grow up and make their own decisions. The LLM has no ability to choose a different life path. The LLM is not getting IDEAS from the book. The LLM is a mathematical engine that will produce what has been asked for, and it will do that by calculating the most likely words to be used based on what has been fed to it.

    The LLM is a machine used to make profit for its programmer, it is not an independent person creating out of inspiration.

    Don't believe the hype. They have NOT produced actual Artificial Intelligence.

  • I think there's an argument that using someone's art or writing to train an AI is like charging for a screening of a movie in your garage. You're using their work and labor for something that will make a profit without their permission. It's not like Fair Use for educational purpose, the AI isn't a human being who can make a choice as to what they do with their education, it's a mathematical prediction engine that is going to be use for industry purposes.

    I can read someone else's book. I can read someone else's book to a child. I can't post someone else's book on my website and charge 5 bucks to read it. I can't reprint someone's book on my website with ads. So why can someone use someone else's book to develop an LLM chatboot that will be placed on a website that gains ad revenue? Or that will be sold to software companies to write technical instructions or code?

    With that in mind, that the lawsuit here is based on COPYING the book to an internal database to train on, based on scanning it, they are arguing that the book was reproduced to gain a profit, basically the same thing as pirating a movie and selling tickets to a private screening.

  • Why the scare quote around fun? He's right. There is no point in adding something to a game that doesn't make it more fun.

  • Privacy is abstract to people until something happens to make them realize how valuable it is. "I don't have anything to hide" is from people who don't feel threatened by anything, who've never been stalked or targeted.

  • When the richest people in the country get loan forgiveness then yes, it can be surprising that the poor don't get it.

  • America, can you explain?

    The system is rigged by the greediest people in the country.

  • The Sage of Mizu continues...