For general use the Framework Desktop fantastic, you can take a look at Geekbench numbers. Or for more comprehensive testing: https://www.phoronix.com/review/ryzen-ai-max-395-9950x-9950x3d/11 - basically it’s neck and neck w/ a desktop 9550X at lower power to boot - really impressive.
Whisper performance on RDNA3 is pretty disappointing. It’ll work, but about on par w/ an RTX 3050 in my testing. I have no idea how a Mac compares for Whisper, however.
For LLMs based on llama2-7b q4_0 llama.cpp benchmarks for Mac vs kyuz0’s Strix Halo tests tg128 (token generation) is about 2X faster on an M3 Ultra, but pp512 (prompt processing) is actually almost exactly the same. You might get some more benefits from running MLX on the Mac.
Personally, if it were me, I’d keep my preorder, but also keep an eye out for if there’s a high memory Mac Mini being released in Oct/Nov (GPU compute on the Mac has been traditionally very underwhelming, but the M5 finally introduces tensor cores), and also if the Nvidia DGX Spark actually finally ships soon as well…
(More memory is definitely better and a good spot to be able to run quants of all the 100B-parameter class MoE models popping up recently.)