Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)LI
Posts
1
Comments
253
Joined
2 yr. ago

  • I don't understand how this will help deep fake and fake news.

    Like, if this post was signed, you would know for sure it was indeed posted by @lily33@lemm.ee, and not by a malicious lemm.ee admin or hacker*. But the signature can't really guarantee the truthfulness of the content. I could make a signed post that claiming that the Earth is flat - or a deep fake video of NASA'a administrator admitting so.

    Maybe I'm missing your point?

    (*) unless the hacker hacked me directly

  • I don't think any kind of "poisoning" actually works. It's well known by now that data quality is more important than data quantity, so nobody just feeds training data in indiscriminately. At best it would hamper some FOSS AI researchers that don't have the resources to curate a dataset.

  • What makes these consumer-oriented models different is that that rather than being trained on raw data, they are trained on synthetic data from pre-existing models. That’s what the “Qwen” or “Llama” parts mean in the name. The 7B model is trained on synthetic data produced by Qwen, so it is effectively a compressed version of Qen. However, neither Qwen nor Llama can “reason,” they do not have an internal monologue.

    You got that backwards. They're other models - qwen or llama - fine-tuned on synthetic data generated by Deepseek-R1. Specifically, reasoning data, so that they can learn some of its reasoning ability.

    But the base model - and so the base capability there - is that of the corresponding qwen or llama model. Calling them "Deepseek-R1-something" doesn't change what they fundamentally are, it's just marketing.

  • Yes, OpenAI wishes everyone else has to have authorization to do model training...

    Fortunately, their ToS don't matter all that much, it's easy to use their model through a third party without ever touching them.

  • The particular AI model this article is talking about is actually openly published for anyone to freely use or modify (fine-tune). There is a barrier in that it requires several hundred gigs of RAM to run, but it is public.

  • It's almost sure to be the case, but nobody has managed to prove it yet.

    Simply being infinite and non-repeating doesn't guarantee that all finite sequences will appear. For example, you could have an infinite non-repeating number that doesn't have any 9s in it. But, as far as numbers go, exceptions like that are very rare, and in almost all (infinite, non-repeating) numbers you'll have all finite sequences appearing.

  • Now, if only the article explained how that killing was related to TikTok. The only relevant thing I saw was,

    had its roots in a confrontation on social media.

    It's says "social media", not "TokTok" though.