Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)IN
Posts
2
Comments
124
Joined
2 yr. ago

  • I've definitely had both. Sometimes the hosts of the actual podcast read an ad using their own voices. In this case everyone gets the same audio file and crowdsourcing the timestamps would work.

    For dynamically inserted ads, it will be more complicated. Maybe a system like content id that has a library of known ads and detects them in the audio.

  • This article is full of errors!

    At its core, an LLM is a big (“large”) list of phrases and sentences

    Definitely not! An LLM is the combination of an architecture and its model parameters. It's just a bunch of numbers, no list of sentences, no database. (Seems like the author confused the word "LLM" with the dataset of the LLM???)

    an LLM is a storage space (“database”) containing as many sample documents as possible

    Nope. This applies to the dataset, not the model. I guess you can argue that memorization happens sometimes, so it might have some features of a database. But it isn't one.

    Additional data (like the topic, mood, tone, source, or any number of other ways to categorize the documents) can be provided

    LLMs are trained in an unsupervised fashion. Just sequences of tokens, no labels.

    Typically, an LLM will cover a single context, e.g. only social media

    I'm not aware of any LLM that does this. What's the "context" of GPT-4?

    software developers have gone to great lengths to collect an unfathomable number of sample texts and meticulously categorize those samples in as many ways as possible

    The closest real thing is the RLHF process that is used to fine tune an existing LLM for a specific application (like ChatGPT). The dataset for the LLM is not annotated or categorized in any way.

    a GPT uses the words and proximity data stored in LLMs

    This is confusing. "GPT" is the architecture of the LLM.

    it is impossible for it to create something never seen before

    This isn't accurate, depending on the temperature setting, an LLM can output literally any word at any time with a non-zero probability. It can absolutely produce things it hasn't seen.

    Also I think it's too simple to just assert that LLMs are not intelligent. It mostly depends on your definition of intelligence and there are lots of philosophical discussions to be had (see also the AI effect).

  • The active users have more than halved since July

    I wouldn't read too much into this, it was a chaotic time, many people tried lots of different things, some created multiple accounts etc. It is completely expected that some try the website and leave again. The growth is still impressive and I expect Lemmy to continue to grow, just because it's the better service.

  • We need something like Affinity for CAD. Surely it must be possible to make a CAD program that can do the basics you need for 3D printing without all the advanced simulations and analysis. I just want to draw some sketches and extrude them.

  • It's so frustrating to see people in this thread posting objectively false statements about SC. Yes, it's behind schedule and yes it suffers from scope creep. But it's not a scam and it's not vaporware. People who give them money know exactly what they are getting into. You can buy a ship now and fly it immediately. You can spend hundreds of hours in the game in it's current state. Even pointing out that it's playable gets downvotes.

  • It would be cool if I could tell it which instance I'm on and then have it change the links to the Lemmy communities so that they are being accessed from my instance. This makes it much easier to interact and subscribe.

    Thank you for this tool!