Read what I said. Intel first put SIMD support on their CPUs in 1999, before GPUs existed as such. They renamed it to AVX in 2008 and widened it to 256 bits, then 512. The fact remains that any software which can benefit from that level of SIMD parallelism can better benefit from it by using the GPU instead, which supports not only vector ops that are effectively millions of bits wide, but tensor ops as well.Or are you really saying Intel put AVX-512 on CPU and then just decided not to use it because, oh, well![]()
Wrong. Main reason why Intel support AVX is latency. Offloading calculations for GPU takes ages latency wise. Therefore CPU is much faster on many cases.Read what I said. Intel first put SIMD support on their CPUs in 1999, before GPUs existed as such. They renamed it to AVX in 2008 and widened it to 256 bits, then 512. The fact remains that any software which can benefit from that level of SIMD parallelism can better benefit from it by using the GPU instead, which supports not only vector ops that are effectively millions of bits wide, but tensor ops as well.
Glad you asked. AVX10 is a simplified version of AVX that drops mandatory 512-bit support:...Also what is AVX10 that Intel will support soon?
You miss the point. When the number of bits you have to process is small, offload latency is indeed a big factor -- but for small workloads, the difference between 256-bit and 512-bit SIMD is small. When you have billions (or trillions) of bits to process, though, offload latency becomes trivial compared to the cost of the operations themselves. I have a software package I run that supports either CPU (via AVX) or GPU modes. The GPU mode is some 5,000% percent faster than the CPU mode, despite the "higher latency".Wrong. Main reason why Intel support AVX is latency. Offloading calculations for GPU takes ages latency wise. Therefore CPU is much faster on many cases.
About whole point of AVX10 is that P-cores can support 512 bits and crap cores less. Right now when crap cores don't support 512 bits, P-cores cannot either. It's not "simplified" but rather offers optional vector length. Something that is missing on current AVX implementations.Glad you asked. AVX10 is a simplified version of AVX that drops mandatory 512-bit support:
"The first and "early" version of AVX10, notated AVX10.1, will not introduce any instructions or encoding features beyond what is already in AVX-512...AVX10.2/256 indicates that a CPU is capable of the second version of AVX10 with a maximum vector width of 256 bits"
Going forward, 512 bit support will only be available on Intel server chips.
For some reason, Zen5 is much faster on AVX512 workloads than Zen4 when using AVX512 supporting software. Not "small" difference. Also for some reason those calculations are not offloaded on GPU. Look at Phoronix 9700X review for more info. 512 bits is right now very possible on desktop too as we can see on Zen5. Transistor budget have not been issue for long time. Zen5 chiplet packs 8 cores under 71mm2. Leaving out L3 cache, cores are very tiny.You miss the point. When the number of bits you have to process is small, offload latency is indeed a big factor -- but for small workloads, the difference between 256-bit and 512-bit SIMD is small. When you have billions (or trillions) of bits to process, though, offload latency becomes trivial compared to the cost of the operations themselves. I have a software package I run that supports either CPU (via AVX) or GPU modes. The GPU mode is some 5,000% percent faster than the CPU mode, despite the "higher latency".
If chip real estate were free, then AVX-1024 or even AVX-8192 would make sense. But every transistor you use for AVX is one less you have in your budget elsewhere.
Intel ditched AVX512 on desktop because their rendition was power hungry and barely anyone was using it.Bwahahaha. Intel only ditched AVX512 because their panic solution has two different type of cores and crappier one does not support AVX512. So to avoid too much compatibility problems, they had to disable it.
Or are you really saying Intel put AVX-512 on CPU and then just decided not to use it because, oh, well![]()
You are confusing Steve conclusion with reality.Here we go again...AMD hype train...overpromised and underdelivered.
![]()
Intel ditched AVX512 on desktop because their rendition was power hungry and barely anyone was using it.
On the other hand, AMD is implementing the same chiplets on desktop and datacenter for their business strategy. In reality, it is a great thing that client users can have access to the same compute performances as the datacenter users.
I literally laughed aloud at this statement. How soon history is forgotten. There were several new CPU releases in the 1980s and early 1990s that saw efficiency leaps of well over 100% -- on any and all workloads, not simply FP performance.incredible value for compute...
Zen 5 might be the most efficient leap in x86 CPUs history.
That's a ballpark larger than any Olympic stadium.Anandtech clearly demonstrated that the IPC gain is in the ballpark of AMD`s claims.
Phoronix results match other reviewers when comparing the same tests, like Blender, code compilation, video encoding with the exception of SVT-AV1, as not every test is using the AVX512 optimizations, and some others. Their average is way different because they test a lot of software that benefits from the changes made on Zen5, the AVX512 improvement in particular pushes it way up.Phoronix has shown how much better these chips are when linux is run on them (so all you linux fans should head over there). Is this then a windows problem much like the high core threadrippers suffered originally?
Zen 5 will go into the worlds most powerfull x86 chips money can buy, beating its main competitor by a wide margin, yeah sure "overpromises" hahHere we go again...AMD hype train...overpromised and underdelivered.
![]()
Read what I said. Intel first put SIMD support on their CPUs in 1999, before GPUs existed as such. They renamed it to AVX in 2008 and widened it to 256 bits, then 512. The fact remains that any software which can benefit from that level of SIMD parallelism can better benefit from it by using the GPU instead, which supports not only vector ops that are effectively millions of bits wide, but tensor ops as well.
The statement is correct; you simply failed to understand it. Hint: look up the term "SIMD", vis a vis MIMD in particular. Any operation a CPU can perform via AVX can indeed be performed by a GPU.Sorry but that's a lie, GPUs can't perform the same tasks as CPUs... let alone with the same latency.
I've already explained above while your statement, though true, has led you to a false conclusion. The hint here is what the "MD" stands for in SIMD. When a CPU executes an command, it does so on one or two operands. Using AVX, that single command operates on 8, 16, or even 32 operands. But with a GPU, that single command can operate on tens of thousands of operands at once....let alone with the same latency. Before the GPU started processing one command, the CPU would have executed dozens.
Nope, it's still inaccurate. You can't, for example, run a JIT recompiler (which uses AVX512) like the one implemented in RPCS3 on a GPU, certain software depends on granularity and ultra low latency.The statement is correct; you simply failed to understand it. Hint: look up the term "SIMD", vis a vis MIMD in particular. Any operation a CPU can perform via AVX can indeed be performed by a GPU.
I've already explained above while your statement, though true, has led you to a false conclusion. The hint here is what the "MD" stands for in SIMD. When a CPU executes an command, it does so on one or two operands. Using AVX, that single command operates on 8, 16, or even 32 operands. But with a GPU, that single command can operate on tens of thousands of operands at once.
Oops! Flatly incorrect. Any instruction capable of being executed with AVX can be executed on a GPU instead. Period. Many JIT compilers do indeed have GPU support:Nope, it's still inaccurate. You can't, for example, run a JIT recompiler (which uses AVX512) on a GPU
That is a disastrous result. Ryzen 1600 > 3600 > 5600 > 7600 were all significant, 20%-ish (give or take) improvements in gaming performance, all while staying in the same 65W TDP. The 9600X being marginally faster than the 7600 in gaming is a big disappointment.
Zen 5 will probably be great for efficiency-sensitive markets like mobile and server, but the desktop chips so far (9600X and 9700X) have little reason to exist.
No, the nature pointed out in these texts is not that of a similar recompiler, a GPU could never RECOMPILE with the latency required for real-time emulation of a complex system. GPUs are good at many tasks, but this is not one of them. It's not that the PS3 team chose it, it's that it's the only way, there isn't a single console emulator that runs JIT recompilers on the GPU. It's kind of crazy to suggest that.Oops! Flatly incorrect. Any instruction capable of being executed with AVX can be executed on a GPU instead. Period. Many JIT compilers do indeed have GPU support:
IBM Java SDK: "You can enable the Just-In-Time (JIT) compiler to offload certain processing tasks to a general-purpose graphics processing unit (GPU). The JIT determines when to offload these tasks based on performance heuristics."
How the JIT compiler uses a GPU (Linux, Windows only)
You can enable the Just-In-Time (JIT) compiler to offload certain processing tasks to a general-purpose graphics processing unit (GPU). The JIT determines when to offload these tasks based on performance heuristics.www.ibm.com
Or this for .NET:
"ILGPU is a JIT (just-in-time) compiler for high-performance GPU programs written in .Net-based languages. It offers the flexibility and the convenience of C++ AMP on the one hand and the high performance of Cuda programs on the other hand...."
How the JIT compiler uses a GPU (Linux, Windows only)
You can enable the Just-In-Time (JIT) compiler to offload certain processing tasks to a general-purpose graphics processing unit (GPU). The JIT determines when to offload these tasks based on performance heuristics.www.ibm.com
The fact that one free, open-source package has chosen to not provide GPU acceleration does not mean it is "impossible".
Yeah sounds like Steve had issues with his chip or maybe BIOS settings or a weird motherboard or something. Funny to see it only slightly ahead of a 5800X3D in many games.
Personally I think if you're on a mostly gaming-focused build using a 5800X3D you might as well wait until AM6/DDR6 to upgrade. Or see if ARM starts taking over from x86 little by little.