Skip Navigation
1 comments
  • The basic model of DeepSeek-R1 14B was already groundbreaking since it reached the level of GPT-1o. But this does much better by bring it to the level of GPT-4o

    Authors are from :
    ::: spoiler 1 - Lightning Rod Labs (USA) ...
    https://www.lightningrod.ai/about :::

    ::: spoiler 2 - (UK) London School of Economics and Political Science
    :::

    Machine learning is still developing very fast.
    "We used 8, H100 GPUs, for training."
    Huge amounts of processing power are not required.