• hendrik
    link
    fedilink
    English
    1
    edit-2
    5 hours ago

    That was my question… How much on-chip memory do they have? And what are applications for that amount of memory? I think an image generator needs like 4-5GB and a LLM that’s smart enough as a general porpose chatbot needs like 8-10GB. More will be better. And at that point you’d better make it unified memory like with the M-series Macs or other APUs? Or this isn’t targeted at generative AI but some other applications. Hence my question.

    • @SlopppyEngineer
      link
      English
      24 hours ago

      Last I heard this is for onboard speech recognition and basic image recognition/OCR so these things can more intelligently listen, see and store what you’re doing without sending it to a server. Not creepy at all.