Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)JS
Posts
14
Comments
1,385
Joined
1 yr. ago

  • Good point. On the other hand, Canada didn't fare any better even without firing its pandemic response team. I suspect project warp speed would have been welcomed either way, even if it was more important in this timeline.

  • I know IPA but I can't read English text written in pure IPA as fast as I can read English text written normally. I think this is the case for almost anyone who has learned the IPA and knows English.

  • I mean, among people who are proficient with IPA, they still struggle to read whole sentences written entirely in IPA. Similarly, people who speak and read chinese struggle to read entire sentences written in pinyin. I'm not saying people can't do it, just that it's much less natural for us (even though it doesn't really seem like it ought to be.)

    I agree that LLMs are not as bright as they look, but my point here is that this particular thing -- their strange inconsistency understanding what letters correspond to the tokens they produce -- specifically shouldn't be taken as evidence for or against LLMs being capable in any other context.

  • When we see LLMs struggling to demonstrate an understanding of what letters are in each of the tokens that it emits or understand a word when there are spaces between each letter, we should compare it to a human struggling to understand a word written in IPA format (/sʌtʃ əz ðɪs/) even though we can understand the word spoken aloud normally perfectly fine.

  • Yes that's right, LLMs are context-free. They don't have internal state. When I say "update on new information" I really mean "when new information is available in its context window, its response takes that into account."

  • Well -- and I don't meant this to be antagonistic -- I agree with everything you've said except for the last sentence where you say "and therefore you're wrong." Look, I'm not saying LLMs function well, or that they're good for society, or anything like that. I'm saying that tokenization errors are really their own thing that are unrelated to other errors LLMs make. If you want to dunk on LLMs then yeah be my guest. I'm just saying that this one type of poor behaviour is unrelated to the other kinds of poor behaviour.

  • as I said, postmodernist lol. I'm coming from the absolutist angle.

    I'll admit though that it also functions to tell you about how someone thinks about the universe. But this is true of any question which has one right answer.

  • in what context? LLMs are extremely good at bridging from natural language to API calls. I dare say it's one of the few use cases that have decisively landed on "yes, this is something LLMs are actually good at." Maybe not five nines of reliability, but language itself doesn't have five nines of reliability.

  • The claim is not that all LLMs are agents, but rather that agents (which incorporate an LLM as one of their key components) are more powerful than an LLM on its own.

    We don't know how far away we are from recursive self-improvement. We might already be there to be honest; how much of the job of an LLM researcher can already be automated? It's unclear if there's some ceiling to what a recursively-improved GPT4.x-w/e can do though; maybe there's a key hypothesis it will never formulate on the quest for self-improvement.