• abhibeckert@beehaw.org
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    1 year ago

    To put some numbers on it - RAM runs at tens of gigabytes per second (bytes, not bits). High Bandwidth Memory runs at several hundred or sometimes terabytes per second (OpenAI is likely using the latter, and that memory isn’t just expensive it’s also supply constrained, so the prices are astronomically high right now).

    You can buy HBM, and you can use it as your main system RAM, but it’s painfully expensive. The actual amount of bandwidth also scales linearly with with the amount of memory you buy as well. So a 500GB is 10x faster than 50GB - because it write to all of the chips simultaneously (and then read from all of them when you access the data back).

    It’s pretty standard on high end GPUs these days. Apple also uses it on all their computers (if you buy a Mac with 64GB of RAM, it’ll run at 800MB/s - which isn’t quite as fast as a high end GPU but it’s close and it is HBM). It’s part of why Macs are so expensive (and also why the cheaper ones have very little RAM).