News:

Willkommen im Notebookcheck.com Forum! Hier können sie über alle unsere Artikel und allgemein über Notebook relevante Dinge disuktieren. Viel Spass!

Main Menu

Post reply

The message has the following error or errors that must be corrected before continuing:
Warning: this topic has not been posted in for at least 120 days.
Unless you're sure you want to reply, please consider starting a new topic.
Other options
Verification:
Please leave this box empty:

Shortcuts: ALT+S post or ALT+P preview

Topic summary

Posted by RobertJasiek
 - September 03, 2023, 08:06:36
Quote from: WhoCares on September 03, 2023, 00:00:54there are also a few tasks nVidia cards can't do at all -

I have never claimed that a) one system would always be better than another system and b) one system could do all tasks for a given expense and time of execution.

Quotefor example, ProRes hardware encoding.

It is well known that hardware encoding is an advantage for specific tasks, such as specific video encoding. Every chip (Apple, Intel, Nvidia etc.) has some such hardware encodings. Unsurprisingly, they help for those specific tasks. However, even so, testers have revealed that it also depends on video en- or decoding settings whether hardware encoding applies and which system is faster for, in particular, video transcoding. In particular, Tech Notice has found that some settings result in relatively much faster, while other settings result in relatively much slower video transcodings of Apple M versus x64 PCs. A statement for ProRes is too general - settings also matter.

QuoteAlso very large neural net model training could be faster on Apple, since some models can't fit in relatively small video memory of consumer nVidia cards.

Could be, or could not be. For it to possibly be faster, software must be written for the used system and its chips.

Neural net is one type of machine learning. Things apply to every kind of machine learning. Indeed, usually training AI requires much more storage (VRAM and RAM, or unified memory) than applying pretrained AI. Indeed, insufficient storage can prevent a task beyond certain application limits.

I apply a pretrained AI on 12 GB VRAM and 64 GB RAM. It can fill 0.8 GB VRAM and 64 GB RAM after ca. 2 hours of execution. I might have spent another €200 to get 128 GB RAM but 64 GB is good enough for me. This exemplifies that it depends on the AI software whether much VRAM, much RAM or both are needed. AAA 3D games are known to sometimes need more than 12 GB VRAM. If I trained the AI I only use, I would profit from gigantic amounts of VRAM (and RAM). How does this compared to 96 GB Apple M unified memory? Within ca. 2 hours, already 64 GB would be filled for RAM-like use so only 32GB would remain for VRAM-like use but to use the latter, more than 64 GB RAM-like use would also be needed. So in practice, 96 GB Apple M unified memory would behave more like 24 GB VRAM + 64 GB RAM, something that is available in a PC as an RTX 4090 (ca. €1650) + 64 GB RAM (ca. €200). Therefore, 96 GB Apple M unified memory  is not at all that impressive except for its lower TDP. For the AI I use, RTX 4070 + 64 GB RAM are 32.5 times faster and 5 times as efficient as Apple M1; Apple M2 96 GB might compare less badly but you get the idea that Nvidia GPU and Nvidia libraries can be much faster and more efficient than Apple M unified memory. I guess that also some AI (video?) software can be found that uses Apple M relatively better, maybe because it also uses some of its hardware transcoding.

It becomes interesting, read: expensive, if we go beyond prosumer Apple M or prosumer PC limitations and demand, say, 1 TB VRAM, 1 TB RAM, 128 core CPU and suitably fast networking / buses. This is territory of €100,000 to €5 million and limitless beyond. Or, more modestly, build an AI workstation with, say, 64 core Threadripper, 8 * RTX 4090, 256 GB RAM for roughly €20,000 ~ €50,000.

For such hardware, AI software design might, or might not, work on distributed storage (especially VRAM) and distributed dGPUs. It often just depends on how that software is designed. There might be specific tasks that cannot work computationally on distributed systems. Usually, it should be possible though. On the software layer, the unified storage model of Apple M can also work when computationally needed. However, recall the 96 GB limit. This advantage becomes meaningless when much more storage is needed and distributed hardware cannot be avoided. I think that, computationally, every algorithm can be transferred from a unified to a distributed approach - it is just a matter of losing some computational speed, but hardly orders of magnitude.

QuoteIf your model can't fit in RAM, your model training will be orders of magnitude slower.

Orders of magnitude slower is really only an issue of VRAM / RAM / unified memory are exceeded and permanent (SSD) storage must be used. Of course, we need enough volatile storage (and speed of the chips).

QuoteMiddle-level Apple laptop has 24-32 Gb of RAM, while high-level Apple laptop/desktop has 64-96 Gb of RAM - and all of it could be used by GPU/Neural Engine part of SoC. Have you ever seen nVidia card with such amount of RAM?

Brainwashing. See above. Just because there is 96 GB unified memory does not mean that 96 GB could be used for either RAM-like or VRAM-like use because the other kind is needed simultaneously.

QuoteDo you know what price is it of?

Much less than an Apple M2 96 GB computer (PC roughly €3000 with RTX 4090 at 24 GB VRAM versus Apple €4500, if my quick Idealo check is right) because PC RAM is cheap. Only the step from 12 to 24 GB VRAM is outrageous (€1000 of the €3000).

For the unlikely split of 48 GB VRAM + 48 GB RAM, the PC with 2 * RTX 4090 will cost ca. €4600, similar to Apple prices. However, I really expect a need for at least 128 GB RAM then (PC €4850 while Apple is impossible with 128 + 48 GB = 176 GB then needed unified memory; maybe Apple M3 for €€€€€?).
Posted by Neenyah
 - September 03, 2023, 00:25:12
You are really refusing to watch the vid, do you? 😀 "Small memory of consumer Nvidia cards" and then an RTX 4070 is far out of reach to the fully maxed M2 Ultra 🙃
Posted by WhoCares
 - September 03, 2023, 00:00:54
>Yes, excellent point Robert 👍

No, his point is not excellent - because there are also a few tasks nVidia cards can't do at all - for example, ProRes hardware encoding. Also very large neural net model training could be faster on Apple, since some models can't fit in relatively small video memory of consumer nVidia cards. If your model can't fit in RAM, your model training will be orders of magnitude slower. Middle-level Apple laptop has 24-32 Gb of RAM, while high-level Apple laptop/desktop has 64-96 Gb of RAM - and all of it could be used by GPU/Neural Engine part of SoC. Have you ever seen nVidia card with such amount of RAM? Do you know what price is it of?
Posted by Neenyah
 - September 02, 2023, 19:38:17
Quote from: RobertJasiek on September 02, 2023, 19:26:18Besides what has already been said, Apple M includes dGPU functionality so must also be compared with x64 PCs with dGPU. There are softwares for which Nvidia dGPU chips are a few times as efficient (application speed per watt) as Apple M2 chips.
Yes, excellent point Robert 👍
Posted by RobertJasiek
 - September 02, 2023, 19:26:18
Quote from: WhoCares on September 02, 2023, 18:30:01M2 Max and M2 Ultra. The latter could be found in Mac Studio desktop and probably can wipe out just anything that Intel or AMD could show in that segment, especially considering TDP.

Besides what has already been said, Apple M includes dGPU functionality so must also be compared with x64 PCs with dGPU. There are softwares for which Nvidia dGPU chips are a few times as efficient (application speed per watt) as Apple M2 chips.

Apple M has its merits but sheep stop trying to brainwash everybody as if it was superior at everything.
Posted by Neenyah
 - September 02, 2023, 19:06:42
Quote from: WhoCares on September 02, 2023, 18:30:01Dude, M2 is base-level laptop SoC, which comes into Apple top-level tablets (iPad Pro), this base-level desktop, base-level laptops (Air 13/15 and 13 Pro)
"Which desktop is right for you?"
( https://imgur.com/I21JRE8 )

Quote from: WhoCares on September 02, 2023, 18:30:01If you still want to wonder, ask yourself how eligible is to compare M2 with 7940HS which is on top of AMD mobile SoCs line, and therefore should be compared with M2 Max.
The M2 Max is 10-11% slower than the 7840H, and that's just a Ryzen 7 ( https://www.cpubenchmark.net/compare/5183vs5558/Apple-M2-Max-12-Core-3680-MHz-vs-AMD-Ryzen-7-7840H ).

Quote from: WhoCares on September 02, 2023, 18:30:01If you really want to get acquainted with real Apple _desktop_ SoC, read about M2 Max and M2 Ultra. The latter could be found in Mac Studio desktop and probably can wipe out just anything that Intel or AMD could show in that segment, especially considering TDP.
Yeah, definitely not. M2 Ultra is being easily beaten by Intel's i7 and i9 and AMD's Ryzen 7 and Ryzen 9 offerings.

"Apple fans, start typing your angry comments now... M2 Ultra Mac Studio Review"
( https://youtu.be/buLyy7x2dcQ )

Watch from 16:07 to the end if you aren't willing to go through the whole vid. You can spend much less to get an overkill PC to destroy that M2 Ultra Mac Studio and still have enough money left to buy one M2 Pro MacBook. Always excellent value at Apple - pay too much to be too much slower at performance per $ 😁

Two very accurate comments under that video:

QuoteI can't believe apple is making the rtx 40 series look like a great value

QuoteOnly Apple could make the Rtx 4090 seem affordable
Posted by WhoCares
 - September 02, 2023, 18:30:01
> you didn't compare this "beast" of a CPU with anything but mobile CPUs

Dude, M2 is base-level laptop SoC, which comes into Apple top-level tablets (iPad Pro), this base-level desktop, base-level laptops (Air 13/15 and 13 Pro). If it is powerful enough to power a desktop - then give cudos to Apple for it, because base-level Intel laptop chips are just not good.

If you still want to wonder, ask yourself how eligible is to compare M2 with 7940HS which is on top of AMD mobile SoCs line, and therefore should be compared with M2 Max.

If you really want to get acquainted with real Apple _desktop_ SoC, read about M2 Max and M2 Ultra. The latter could be found in Mac Studio desktop and probably can wipe out just anything that Intel or AMD could show in that segment, especially considering TDP.
Posted by Neenyah
 - September 02, 2023, 15:10:28
Funny how, despite calling this Mini "desktop" 11 times in the article, you didn't compare this "beast" of a CPU with anything but mobile CPUs. And then it still got destroyed by those. Apple FTW.
 
Edit: This is why there is no comparison with desktops:
"PC vs MAC for the SAME PRICE - Which is better for VIDEO exporting?"
( https://youtu.be/_D0K3-uZMyY )
 
Desktop PC at the same price is wiping the floor with both M2 and M2 Pro, lol, with just i5 13600K + RTX 3050 as shown in that video.

Funny quote from Apple's site: https://www.apple.com/mac-mini/

QuoteWhether you choose M2 or M2 Pro, the performance and efficiency of Apple silicon allow Mac mini to blow away desktops many times its size — all in an iconic 7.7‑inch‑square frame.

😄
Posted by Andreas Galster
 - September 02, 2023, 11:34:07
The new M3 Mac Mini is supposed to be announced next month... A very timely review...?! lol
Posted by Redaktion
 - September 02, 2023, 09:54:10
Apple's Mac Mini now includes the latest M2 and M2 Pro CPUs, delivering plenty of power for a compact desktop. Since upgrades are costly, like they are for MacBooks, the small device can get pricey.

https://www.notebookcheck.net/Apple-Mac-Mini-M2-2023-review-Apple-M2-unleashing-its-power-via-desktop.745320.0.html