• @A_A
    link
    English
    3
    edit-2
    3 days ago

    The basic model of DeepSeek-R1 14B was already groundbreaking since it reached the level of GPT-1o. But this does much better by bring it to the level of GPT-4o

    Authors are from :

    1 - Lightning Rod Labs (USA)


    https://www.lightningrod.ai/about

    2 - (UK)

    London School of Economics and Political Science


    Machine learning is still developing very fast.
    “We used 8, H100 GPUs, for training.”
    Huge amounts of processing power are not required.