• 0 Posts
  • 31 Comments
Joined 11 months ago
cake
Cake day: March 22nd, 2024

help-circle



  • Most CUDA or PyTorch apps can be run through ROCM. Your performance/experience may vary. ZLUDA is also being revived as an alternate route to CUDA compat, as the vast majority of development/intertia is with CUDA.

    Vulkan has become a popular “community” GPU agnostic API, all but supplanting OpenCL, even though it’s not built for that at all. Hardware support is just so much better, I suppose.

    There are some other efforts trying to take off, like MLIR-based frameworks (with Mojo being a popular example), Apache TVM (with MLC-LLM being a prominent user), XLA or whatever Google is calling it now, but honestly getting away from CUDA is really hard. It doesn’t help that Intel’s unification effort is kinda failing because they keep dropping the ball on the hardware side.



  • And even better, “testing” it. Maybe I’m sloppy, but I have failed runs, errors, hacks, hours of “tinkering,” optimizing, or just trying to get something to launch that feels like an utter waste of an A100 mostly sitting idle… Hence I often don’t do it at all.

    One thing you should keep in mind is that the compute power of this thing is not like an A/H100, especially if you get a big slowdown with rocm, so what could take you 2-3 days could take over a week. It’d be nice if framework sold a cheap MI300A, but… shrug.




  • The APU/RAM is one unit, everything else is modular and repairable. They aren’t price gouging the RAM either.

    They’re one upping Apple, big PC OEMs and Chinese Mini PC makers with a more repairable, consumer friendly product. That sounds like Framework to me.

    I think you are clinging to the idea that RAM will be separate and upgradable from the CPU for a long time… Physics dictates that it will not, especially in the laptop space where wasted power is so critical. Hacks needed to even make that work now with DDR5 are kinda crazy and inefficient (just look at the voltage/speeds/timings ddr5 sodimms run at). LPCAMM is supposedly a good stopgap, but even that is having teething issues