Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)LW
Posts
19
Comments
1,833
Joined
2 yr. ago

  • OpenAI has gotten virtually unlimited funding for years. It has first dibs and deep discounts on Microsoft data centers.

    And somehow, despite every single trade restriction, multiple random startup companies in China (that don't even know how to secure their own databases) manage to make LLMs that outperform it.

    I'm not saying that because Chinese companies are uniquely cool. I'm saying that because this whole AI thing is uniquely stupid.

  • I figured as much. Even this line...

    M1's capabilities are top-tier among open-source models

    ... is right above a chart that calls it "open-weight".

    I dislike the conflation of terms that the OSI has helped legitimize. Up until LLMs, nobody called binary blobs "open-source" just because they were compiled using open-source tooling. That would be ridiculous

  • I've been able to run distillations of Deepseek R1 up to 70B

    Where do you find those?

    There is a version of Deepseek R1 "patched" with western values called R1-1776 that will answer topics censored by the Chinese government, however.

    Thank you for mentioning this, as I finally confronted my own preconceptions and actually found an article by Perplexity that demonstrated R1 itself has demonstrable pro-China bias.

    Although Perplexity's own description should cause anybody who understands the nature of LLMs to pause. They describe it in their header as a

    version of the DeepSeek-R1 model that has been post-trained to provide unbiased, accurate, and factual information.

    That's a bold (read: bullshit) statement, considering the only altered its biases on China. I wouldn't consider the original model to be unbiased either, but apparently perplexity is giving them a pass on everything else. I guess it's part of the grand corporate lie that claims "AI is unbiased," a delusion that perplexity needs to maintain.

  • That's because search engines have reached the stage of enshittification where they no longer need to be good. Instead, they want you to spend as much time there as possible.

    LLMs are still being sold as "the better option" - including by the exact same search giants who intentionally ruined their own search results. And many of them are already prioritizing agreeableness over "truthfulness." And we're still in the LLM honeymoon phase, where companies are losing billions of dollars on a yearly basis and undercharging their users.

  • If you're talking about the distillations, AFAIK they take somebody else's model and run it through their (actually open-source) distiller. I tried a couple of those models because I was curious. The distilled Qwen model is cagey about Tianmen Square, but Qwen was made by Alibaba. The distillation of a US-made model did not have this problem.

    (Edit: we're talking about these distillations, right? If somebody else ran a test and posted it online, I'm not privy to it.)

    I don't have enough RAM to run the full DeepSeek R1, but AFAIK it doesn't have this problem. Maybe it does.

    In case it isn't clear, BTW, I do despise LLMs and AI in general. The biggest issue with their lies (leaving aside every other issue with them for a moment) isn't the glaringly obvious stuff. Not Tianmen Square, and certainly not the "it's woke!" complaints about generating images of black founding fathers. The worst lies are the subtle and insidious little details like agreeableness - trying to get people to spend a little more time with them, which apparently turns once-reasonable people into members of micro-cults. Like cults, perhaps, spme skeptics think they can join in and not fall for the BS... And then they do.

    All four students had by now joined their chosen groups... Hugh had completely disappeared into a nine-week Arica training seminar; he was incommunicado and had mumbled something before he left about “how my energy has moved beyond academia.”

  • What exactly makes this more "open source" than DeepSeek? The linked page doesn't make that particularly clear.

    DeepSeek doesn't release their training data (but they release a hell of a lot of other stuff), and I think that's about as "open" as these companies can get before they risk running afoul of copyright issues. Since you can't compile the model from scratch, it's not really open source. It's just freeware. But that's true for both models, as far as I can tell.

  • DeepSeek imposes similar restrictions, but only on their website. You can self-host and then enjoy relatively truthful (as truthful as a bullshit generator can be) answers about both Tianmen Square, Palestine, and South Africa (something American-made bullshit generators apparently like making up, to appease their corporate overlords or conspiracy theorists respectively).

  • Edit:

    this site is mostly ads now

    • Only 4 of the original 22 options are still on the homepage
    • Two (50%) are hosted via sponsor CloudBreak
    • One (25%) is hosted via sponsor Sunya
    • The "get started" links mostly (75%) point to Stripe checkout pages

    OP, is this a joke?


    Sponsored, using affiliate links and accepting donations? Somebody better fork this guide before the GitHub gets yanked

    Edit: Okay, after looking around at this, something seems... off. Linking to getoffpocket.com?by=lemmy was odd already, but then I noticed that every single service here appears to have a referral link. Even the OneNote link has a referral code stapled onto it:

     
        
    https://www.microsoft.com/en-us/microsoft-365/onenote/digital-note-taking-app?rby=getoffpocket.com%2Fproprietary%2Fmicrosoft-onenote%2F
    
      

    For some reason, those same UTM links are used for everything, including links to GitHub?
    How about no extra query parameters at all?

    I'm also surprised there's not even a passing mention to Obsidian and Evernote.

    I think I'll stick to searching out my own recs on AlternativeTo