News:

Willkommen im Notebookcheck.com Forum! Hier können sie über alle unsere Artikel und allgemein über Notebook relevante Dinge disuktieren. Viel Spass!

Main Menu

Post reply

The message has the following error or errors that must be corrected before continuing:
Warning: this topic has not been posted in for at least 120 days.
Unless you're sure you want to reply, please consider starting a new topic.
Other options
Verification:
Please leave this box empty:

Shortcuts: ALT+S post or ALT+P preview

Topic summary

Posted by Obero
 - December 24, 2020, 19:59:48
So AMD is going to give tough time to intel
Posted by Vyor
 - December 22, 2020, 17:51:03
Quote from: Ben3000 on December 22, 2020, 03:39:18
Quote from: Mate on December 20, 2020, 22:22:16
@_MT_
(sorry for links, I cant post with those im comment :< )
QuoteIf you need to run at 4.5 GHz to match the performance, I can't see you matching efficiency (under high load) on the same manufacturing node.

Even if M1 high performance cores are probably a lot larger? As I said earlier my Ryzen 4900HS clearly beats M1 if I underclock(underTDP? dunno how to describe it) it to 20W, so in theory it should be under load as efficient as M1 sucking 14W. Anandtech reviewed m1 mac and discorevered that under one core load mini is draining 10.5W while on idle its only 4.2W.
images.anandtech. com /graphs/graph16252/119344.png


Let me quickly dispel something. Underclocking is the same as overclocking. You are trying to get as close as possible to the edge of the capacity of the silica.

There is a reason why your 4900HS uses a specific amount of power, in a specific range. Namely, to enable as much as possible CPU's to be sold, within a specific range. So your comparison is the same as overlocking a CPU and stating its faster then the competition, while maybe 90% of the other owners of that same CPU, can not get that gain.

Next error:

You are looking at 10.5W WALL drawing power for the single core performance of the mac Mini. That means:

* Power conversion loss
* Memory power consumption
* GPU
* Controller
* Fan
* ... and every other component

Where as the numbers you are trying to use for your 4900HS, are totally different measurements.

youtu+++be/y_jw38QD5qY?t=684

Replace the +++ with a dot.

Notice how in real world, the package power reported by the 4900HS is 54W, then drops to 35W for 20% faster performance. While the M1 stays at 13W, for 20% less performance then the 4900HS, with 2.4 to 4.8 times less power used.

A theoretical 8+4 Core M1X can produce 30% faster results then a 5900HS at twice the power efficiency given the benchmark leaks. So no, AMD can not magically make power / efficiency appear out of thin air.

AMD and Intel produce desktop CPUs, where power consumption is secondary to performance. And then downscale those to Laptops.

Apple used Phone parts, where power efficiency is primary and then upscaled them to their laptop parts. And this approach is paying off much better in efficiency.

And while one can complain about Apple 5nm. Apple did the investment to buy 5nm space ( and 80% 5nm in 2021 ). AMD did not. AMD people did not complain when AMD got 7nm and Intel was still stuck on 14nm.

So the whole "things will be different when AMD is at 5nm". By that time Apple we are looking past 2022 ( AMD is going 6nm in 2022 based upon the leaks ). That means Apple is already looking at the M3/M3X + 3nm.

People underestimate how much of a leap Apple has on the competition in this regard. Anyway, enjoy the fun competition times. One thing i see is that for Apple, it was a good move to do away with Intel, as a lot more people are interested in their products as a result.

It wasn't wall power, it was from Apple's internal metrics.

Go away ARM fanboy
Posted by Ben3000
 - December 22, 2020, 03:39:18
Quote from: Mate on December 20, 2020, 22:22:16
@_MT_
(sorry for links, I cant post with those im comment :< )
QuoteIf you need to run at 4.5 GHz to match the performance, I can't see you matching efficiency (under high load) on the same manufacturing node.

Even if M1 high performance cores are probably a lot larger? As I said earlier my Ryzen 4900HS clearly beats M1 if I underclock(underTDP? dunno how to describe it) it to 20W, so in theory it should be under load as efficient as M1 sucking 14W. Anandtech reviewed m1 mac and discorevered that under one core load mini is draining 10.5W while on idle its only 4.2W.
images.anandtech. com /graphs/graph16252/119344.png


Let me quickly dispel something. Underclocking is the same as overclocking. You are trying to get as close as possible to the edge of the capacity of the silica.

There is a reason why your 4900HS uses a specific amount of power, in a specific range. Namely, to enable as much as possible CPU's to be sold, within a specific range. So your comparison is the same as overlocking a CPU and stating its faster then the competition, while maybe 90% of the other owners of that same CPU, can not get that gain.

Next error:

You are looking at 10.5W WALL drawing power for the single core performance of the mac Mini. That means:

* Power conversion loss
* Memory power consumption
* GPU
* Controller
* Fan
* ... and every other component

Where as the numbers you are trying to use for your 4900HS, are totally different measurements.

youtu+++be/y_jw38QD5qY?t=684

Replace the +++ with a dot.

Notice how in real world, the package power reported by the 4900HS is 54W, then drops to 35W for 20% faster performance. While the M1 stays at 13W, for 20% less performance then the 4900HS, with 2.4 to 4.8 times less power used.

A theoretical 8+4 Core M1X can produce 30% faster results then a 5900HS at twice the power efficiency given the benchmark leaks. So no, AMD can not magically make power / efficiency appear out of thin air.

AMD and Intel produce desktop CPUs, where power consumption is secondary to performance. And then downscale those to Laptops.

Apple used Phone parts, where power efficiency is primary and then upscaled them to their laptop parts. And this approach is paying off much better in efficiency.

And while one can complain about Apple 5nm. Apple did the investment to buy 5nm space ( and 80% 5nm in 2021 ). AMD did not. AMD people did not complain when AMD got 7nm and Intel was still stuck on 14nm.

So the whole "things will be different when AMD is at 5nm". By that time Apple we are looking past 2022 ( AMD is going 6nm in 2022 based upon the leaks ). That means Apple is already looking at the M3/M3X + 3nm.

People underestimate how much of a leap Apple has on the competition in this regard. Anyway, enjoy the fun competition times. One thing i see is that for Apple, it was a good move to do away with Intel, as a lot more people are interested in their products as a result.
Posted by Vyor
 - December 21, 2020, 15:28:03
Quote from: xpclient on December 21, 2020, 14:09:47
Quote from: vertigo on December 20, 2020, 18:57:49
While GPU encoding is indeed much faster, it's nowhere near as good, and it results in worse quality and larger file sizes. I know that's changed a lot with the RTX NVENC, but even though it's a lot better, I still wonder if it can compete with software encoding on pure quality, i.e. it might be comparable at similar speeds, but I suspect software encoding, when slowed down, would still be superior. So if you want to encode videos, software encoding is still probably the best option in order to maximize quality and file size, whereas if you're talking about Twitch streaming, yeah, NVENC is probably better, or at least as good.
It can compete on pure quality according to this comparison: REDACTED LINK It beats x264 on Slow as well. Every encoding task has to be evaluated in terms of time spent too. And Turing NVENC sort of marked a transition where the quality being so good and the speed insanely fast, the only reason left now to use software encoding is if you want to use low bitrates (2000 kbits/s or lower). Everyone's connections too are fast enough to stream high bitrate video.

Twitch does not accept HEVC or h.265. Medium x.264 beats everything else.

For recording, x.265 medium beats all else bar AV1, which is way too heavy on things to use.
Posted by xpclient
 - December 21, 2020, 14:09:47
Quote from: vertigo on December 20, 2020, 18:57:49
While GPU encoding is indeed much faster, it's nowhere near as good, and it results in worse quality and larger file sizes. I know that's changed a lot with the RTX NVENC, but even though it's a lot better, I still wonder if it can compete with software encoding on pure quality, i.e. it might be comparable at similar speeds, but I suspect software encoding, when slowed down, would still be superior. So if you want to encode videos, software encoding is still probably the best option in order to maximize quality and file size, whereas if you're talking about Twitch streaming, yeah, NVENC is probably better, or at least as good.
It can compete on pure quality according to this comparison: https://unrealaussies.com/tech/nvenc-x264-quicksync-qsv-vp9-av1/ It beats x264 on Slow as well. Every encoding task has to be evaluated in terms of time spent too. And Turing NVENC sort of marked a transition where the quality being so good and the speed insanely fast, the only reason left now to use software encoding is if you want to use low bitrates (2000 kbits/s or lower). Everyone's connections too are fast enough to stream high bitrate video.
Posted by Vyor
 - December 20, 2020, 23:51:38
Quote from: Mate on December 20, 2020, 22:45:08
QuoteImagine that instead of having AVX512 units inside your generic high performance cores, you'd have a core specifically designed for SIMD which could go even wider than 512. Yes, it's the iGPU
Actually GPU is also coprocessor, same thing with sound card on mother board

  I think Mate is a bit confused on this. x86 does have certain properties which complicate implementation. And because of backward compatibility, they can't get rid of them.
I was  trying to say exactly this. x86 decoders are a lot more complex and  now you want to add additional tasks on them too.

Have you guys never heard of the "fusion" initiative by AMD, something that started after they purchased ATI?

They found that shoving things onto their own coprocessor, in this case an iGPU to handle FP tasks, was a really bad idea for most things.

Not only do you have massive instruction latency, you also have a massive power hog in moving the data off core into the accelerator. Then you have the issue of getting devs to actually use your special instructions, not easy, and then making sure to support those instructions for more or less every product you make with that capability(don't want someone to buy an upgrade to a part to get better AI perf and then them finding out you changed the instruction so their software no longer works).

And no, ARM is not immune to that last problem. ARM instructions go a very, very long time before being depreciated, ARMv8 shares a lot of instructions in common with even ARMv5.

And, again, it isn't really that hard to do in the first place, both AMD and Intel have done and will continue to do it for things like AI and video decode/encode, it just takes dev support; something harder to acquire than gold out of a duck's a**.

Quote
I'm certain it was one of key factors of CISC success in 80s when RAM was very small. Now it doesnt make almost any impact. Code occupying 2x more space in memory? No problem at all when simple GUI takes a lot more.

Registers and instruction caches are still really small dude. That's what the variable length is for. The 4800U actually has a smaller instruction cache than the athlon XPs did, half the size in fact.

The M1, meanwhile, has an absolutely humongous 192kb L1 instruction cache, this compared to Zen3's 32kb L1 instruction cache. And this cache is power hungry too, most of the power a CPU takes can be divided into three parts:

FP Units(now the biggest factor)
Cache(second biggest factor)
Front End(this includes decode, though that's very rapidly ceased to be a major component in things, partially thanks to decoder width staying very static over time, the Core architecture only lacked 1 decoder compared to Skylake, ditto K8 compared to Zen3)

The more cache you have, the more power hungry your part will be by default, regardless of anything else. It is unbelievably hard to clock gate cache, and the more you have the harder it becomes.

This is why x86-64 and even Itanium/Intel64 stuck with variable length instructions. Them being CISC just means variable length actually nets them a bit of decoding ease on top.

QuoteAgree - this is huge problem. x86 processors needs to run even code prepared for 286.... on the other hand ARM have only 32/64 switch, not compatibility with 16/32/64 and dozens of instruction sets extensions.

Modern x86 CPUs do not have native compatibility for 16 bit code, it's all emulated. That's why the x87 instruction set is so god damn slow on them. It's very similar to ARM in that way, older ISAs just get emulated and depreciated.
Posted by Mate
 - December 20, 2020, 22:45:08
QuoteImagine that instead of having AVX512 units inside your generic high performance cores, you'd have a core specifically designed for SIMD which could go even wider than 512. Yes, it's the iGPU
Actually GPU is also coprocessor, same thing with sound card on mother board

  I think Mate is a bit confused on this. x86 does have certain properties which complicate implementation. And because of backward compatibility, they can't get rid of them.
I was  trying to say exactly this. x86 decoders are a lot more complex and  now you want to add additional tasks on them too.

QuoteFor example, perhaps exactly because x86 had many specialized instructions, they opted for variable length (it allows you to make common instructions short so you can get away with smaller instruction cache and you use less memory and storage; you can take less space on average while having a lot more instructions


I'm certain it was one of key factors of CISC success in 80s when RAM was very small. Now it doesnt make almost any impact. Code occupying 2x more space in memory? No problem at all when simple GUI takes a lot more.

QuoteOne big advantage of ARM is that they're not afraid to break backward compatibility. Intel has to live with decisions they made in a different age.
Agree - this is huge problem. x86 processors needs to run even code prepared for 286.... on the other hand ARM have only 32/64 switch, not compatibility with 16/32/64 and dozens of instruction sets extensions.
Posted by _MT_
 - December 20, 2020, 22:33:23
Quote from: vertigo on December 20, 2020, 17:05:43
That makes sense. Thanks. Still seems like they're playing some misleading marketing trickery, though, trying to play it off as though it's several generations in. Technically, the iGPU is, but the Xe design isn't.
I think it's silly since I doubt the iGPU actually has 12 generations (that they really improved it every single time). On the other hand, I think that it's more consumer friendly and avoids unnecessary confusion. It's a processor. The GPU is inside. Having multiple generations associated with a single product might give information to us, but it would just confuse ordinary consumers. Consider the mess AMD created with Zen+, releasing mobile 3000 series with a different core then desktop, skipping 4000 series on desktop and then mixing two cores in mobile 5000 series. What a dog's breakfast.
Posted by Mate
 - December 20, 2020, 22:22:16
@_MT_
(sorry for links, I cant post with those im comment :< )
QuoteIf you need to run at 4.5 GHz to match the performance, I can't see you matching efficiency (under high load) on the same manufacturing node.

Even if M1 high performance cores are probably a lot larger? As I said earlier my Ryzen 4900HS clearly beats M1 if I underclock(underTDP? dunno how to describe it) it to 20W, so in theory it should be under load as efficient as M1 sucking 14W. Anandtech reviewed m1 mac and discorevered that under one core load mini is draining 10.5W while on idle its only 4.2W.
images.anandtech. com /graphs/graph16252/119344.png

So we have 5-6W for Firestorm core under load - its a lot. If we do math 5*8  = 40W
next.lab501. ro/notebook/english-lenovo-ideapad-s540-13are-vs-13iml-amd-ryzen-7-4800u-vs-intel-core-i7-10710u/13
You can check there that with 42W power(~5,6W for each core) is needed for 4800U to boost all cores to 4300mhz. Yeah I know we cant simplify things like this but it shows that  difference in efficiency under load is smaller than you would expect. What makes difference is BIG.little as M1 can turn off and on cores when needed. Also way how x86 CPUs work doesnt help, as especially mobile CPUs relies on short burtst to make things smoother for user. However power consumptions rises expotentially with higher clocks.

QuoteYes, Apple has manufacturing edge. But that core kicks a**. So much that it can bear comparison with Zen 3 running at desktop frequencies. Can you imagine what 8 Firestorm cores could do with a decent power limit?

Ok, so lets do fast math. M1 in Cinebench - singlecore 1514, multicore 7760. So 4x Firestorm are responsible for ~6000cb, Icestorm adds ~1700. 8 x 1500(8x Firestorm) + 1700 = 13800. We also know from Anandtech that it should result in around 45W energy consumption. Cezanne-H will probably score around 12000-13000 if we consider slightly improved manufacturing process and higher IPC. Yes, thats lower score, but pretty impressive given that M1 has 30% energy consumption handicap.

Anyway I'm not here to defend AMD - true revolution that Apple brings is in coprocessors that will allow to do many common tasks with significantly less effort from CPU.



@Vertigo

QuoteHopefully, though, the M1 will push AMD and Intel to improve their efficiency, so x86 options will become better with battery life and noise as well,
They would need to tweak CPUs to work on lower TDP and with significantly reduced boost clock. Also manufacturing process is bigger 'player' there. Before Renoir AMD was using 12nm process that was worse than Intel 14nm, their mobile CPUs sucked and on top of that Intel had significant edge in battery life too. AMD celebrated switch to 7nm by releasing CPUs with 2x better performance than generation before and with better energy efficiency.

QuoteSounds like an interesting approach, but the downside would be that due to their specialization those coprocessors can only be utilitized a portion of the time, vs the CPU being able to be used for everything, so while it would be more efficient, it will require more space.

Thats why you also need good general performance of CPU. Space is also no problem if you are using industry leading manufacturing process. Key point of those coprocessors is that specialized circuits can do their job immensely more efficient. For example i7-4870HQ is significantly more powerful than i3-1000NG4 in general tasks. However if you try playing 4k video on youtube then i7 will turn fans very fast and will be sweating- even then it will drop a lot of frames. in same time i3 will be chilling and its video encoding/decoding coprocessor  will handle everything smoothly.   Insanely efficient coprocessors are clearly future - you can add only handful ones that improve performance in most popular areas.
Posted by _MT_
 - December 20, 2020, 22:18:16
Quote from: vertigo on December 20, 2020, 19:19:02
If I'm understanding you correctly, you're saying ARM, due to its relative simplicity, is easier to implement coprocessors, and instead of having one powerful CPU doing everything, they're working toward a central CPU (yes, I realize that's redundant, but intentionally so) for main tasks that's assisted by numerous, specialized coprocessors for various different tasks. So basically, using delegation to accomplish things with better efficiency vs just having everything done by a more generalized chip (the CPU). Is that right? Sounds like an interesting approach, but the downside would be that due to their specialization those coprocessors can only be utilitized a portion of the time, vs the CPU being able to be used for everything, so while it would be more efficient, it will require more space.
I think I would call what they're doing a unified processor. A processor that offers a number of different cores with very different characteristics that can cooperate efficiently (instead of number of separate processors that are external to each other). Imagine that instead of having AVX512 units inside your generic high performance cores, you'd have a core specifically designed for SIMD which could go even wider than 512. Yes, it's the iGPU. Why not look at iGPU units as just another type of core that is very good at certain types of tasks. Instead of having one monster core that can do everything, let's have independent cores that can each do very different things (it's, really, a question of packaging, organization). What I'm curious about is how far can they take it. When they run out of die space, what will they do. It's one thing to do this for MBA. It's another to scale this up to Mac Pro. M1 has four types of cores inside. Historically, iGPUs were too weak for such thoughts. They existed for efficiency, not performance. The problem with dGPU is that it's far away and there are significant overheads (it's worth it only for relatively big tasks). I think they're looking in this direction - improving cooperation through better integration.

Modern x86 processors don't work as x86 internally. You could say they're just emulating x86, in hardware. And having specialized instructions is a CISC philosophy (minimizing the total number of cycles needed to perform a task), it's exactly what x86 did. RISC processors learned to do this as well. The thing is that in some cases, x86 processors had enough generic performance to not need acceleration that weak ARM processors needed in order to deliver decent user experience (which can then result in a weak processor matching or outperforming a more powerful processor - you have to be careful when comparing them). I think Mate is a bit confused on this. x86 does have certain properties which complicate implementation. And because of backward compatibility, they can't get rid of them. For example, perhaps exactly because x86 had many specialized instructions, they opted for variable length (it allows you to make common instructions short so you can get away with smaller instruction cache and you use less memory and storage; you can take less space on average while having a lot more instructions). But now, variable length complicates parallel decoding. One big advantage of ARM is that they're not afraid to break backward compatibility. Intel has to live with decisions they made in a different age.
Posted by _MT_
 - December 20, 2020, 20:25:34
Quote from: Mate on December 20, 2020, 18:26:17
Well, it depends how you define efficiency. According to TSMC 5nm improved energy efficiency by 30% and M1 is using 14W under load.  Ryzen Renoir(8 cores) can easily outperform Apple chip in general tasks with 20W so in theory it should be more efficient under load as 20*0,7 = 14.
I certainly wouldn't put it that way ("can easily outperform"). I define efficiency as an average number of operations per watt in a given load (and I'm interested in general arithmetics). You have to remember that M1 is a 4+4 configuration. There are two different cores at play. It's the high performance, Firestorm cores that are of interest to me when we are comparing the state of technology. If memory serves me right, Firestorm running at around 3 GHz can square up with Zen 3 running at roughly 4.5 GHz (Zen 2 can't go high enough). That's a huge difference and it makes a significant difference to efficiency (it also depends on where exactly the voltage-frequency curve sits). The higher the frequency, the higher the voltage to maintain stable operation. And power goes up with the square of voltage (meaning efficiency can drop quickly at higher frequencies). If you need to run at 4.5 GHz to match the performance, I can't see you matching efficiency (under high load) on the same manufacturing node.

Sure, AMD has processors with 8 full-fat cores with SMT (and even more for desktops). Apple currently doesn't. Firestorm, AFAIK, doesn't even have SMT. Which only makes its performance even more impressive. Kind of like with Intel, AMD benefits from having more ("big") cores which allows them to run at lower frequencies (where the cores are more efficient) than they would have to if the opponent had the same configuration. But that's just packaging. To me, that's noise. Something to be removed. It matters when you're buying a product. It doesn't matter when you're comparing the underlying technology.

It's not apples to apples when you compare M1 against something like 4800U. In a way, M1 is more like a 4 core processor with SMT; each big core just has an additional smaller core instead of SMT. And you wouldn't expect a 4C/8T processor to compete against an 8C/8T processor just because they support the same number of hardware threads. Consider that M1 performs roughly like 4700U. Which also doesn't have SMT. So, you've got 4 Firestorm + 4 Icestorm cores matching 8 Zen 2 cores with SMT disabled. While pulling less power. Yes, Apple has manufacturing edge. But that core kicks a**. So much that it can bear comparison with Zen 3 running at desktop frequencies. Can you imagine what 8 Firestorm cores could do with a decent power limit?
Posted by vertigo
 - December 20, 2020, 19:19:02
Quote from: Mate on December 20, 2020, 18:26:17
@_MT_
QuoteEven if AMD was currently on the 5nm node, they wouldn't be able to match M1's efficiency.
Well, it depends how you define efficiency. According to TSMC 5nm improved energy efficiency by 30% and M1 is using 14W under load.  Ryzen Renoir(8 cores) can easily outperform Apple chip in general tasks with 20W so in theory it should be more efficient under load as 20*0,7 = 14.

However its clear that Apple aims to cover many use cases with dedicated coprocessors  that will make their processors very efficient. It is also probably easier to implement more coprocoessors for ARM  than in x86 as this makes x86 decoders even more complex. As things are now I can clearly see Macbooks in year or two just as better alternative for x86 premium laptops in performance/price ratio in every segment. AMD and Intel have very complex processor designs and it will be harder for them to keep up with  Apple each year.

@Vertigo
Maybe I used wrong words - I'm using Windows(for my personal computer) and MacOS side by side for years. However now M1 is very close to meet all my performance expectations(also for gaming!) and blow away x86 devices in terms of battery life and silence. My G14 would be already on ebay if m1 macbooks could  handle two external displays  in normal way. I'm really tired of that annoying noise and heating for no reason.

Yeah, that makes more sense. Not so much looking to switch as drop one.  My point was that you should go with whichever one suits your needs better, regardless of the hardware, but obviously they both work for you, and you're familiar with both. Personally, I don't care for the Apple ecosystem, so as much as I hate Windows, that's what I use because it does what I need, even if it's a major PITA in the process (I'd rather go with Linux, but I've tried it multiple times and it just doesn't work for me). Hopefully, though, the M1 will push AMD and Intel to improve their efficiency, so x86 options will become better with battery life and noise as well, and with MS exploring ARM now as well, it seems that may become an option outside of Macs in the hopefully not too distant future.

As for gaming, yeah, the M1's performance there is very impressive, but there's still the issue of many/most games not working, or not working well, on Mac. And now you can't even dual-boot Windows, though from what I've seen it emulates well. I'm sure you're following that much more closely than I am, but worth mentioning.

If I'm understanding you correctly, you're saying ARM, due to its relative simplicity, is easier to implement coprocessors, and instead of having one powerful CPU doing everything, they're working toward a central CPU (yes, I realize that's redundant, but intentionally so) for main tasks that's assisted by numerous, specialized coprocessors for various different tasks. So basically, using delegation to accomplish things with better efficiency vs just having everything done by a more generalized chip (the CPU). Is that right? Sounds like an interesting approach, but the downside would be that due to their specialization those coprocessors can only be utilitized a portion of the time, vs the CPU being able to be used for everything, so while it would be more efficient, it will require more space.
Posted by vertigo
 - December 20, 2020, 18:57:49
Quote from: xpclient on December 20, 2020, 17:52:28
Quote from: neblogai on December 20, 2020, 14:12:25
RDNA2 is coming to Van Gogh- a competitor to Intel's UP4 APUs, and it will likely trounce it in games in ultraportable/fanless. For gaming laptops- H-series will use dGPU, U-series can use dGPUs like MX350/450 too. And that iGPU in TigerLake is nothing special- Intel was showing wonders their slides, but in real life laptop tests- even current Renoir with Vega iGPU goes toe to toe with them. Lucienne/Cezanne, with higher enabled clocks and numbers of CUs, is likely to be somewhat faster.
The only real area where Intel leads, is Thunderbolt. Cezanne is still PCIe 3.0, and no thunderbolt or USB4. So for external dGPUs- Intel is the only option.

You are thinking purely from a gaming performance perspective.  I was thinking more from a media encoding perspective, not just gaming. I already have a gaming laptop with Intel Coffee Lake Refresh CPU+Nvidia RTX 2080 GPU but I also use it heavily for media encoding since NVENC is the best in the industry there is for H.264 and HEVC encoding after Turing.

Tiger Lake has added VP9 so I'd like to check it out, and a reworked quality-focused HEVC encoder and AV1 hardware decoding which makes it attractive for this market.  :) Nvidia's MX graphics lack NVENC entirely and AMD's Video Core Next is lacking in quality and features compared to Intel and Nvidia's implementations.

You may already know this, but while GPU encoding is indeed much faster, it's nowhere near as good, and it results in worse quality and larger file sizes. I know that's changed a lot with the RTX NVENC, but even though it's a lot better, I still wonder if it can compete with software encoding on pure quality, i.e. it might be comparable at similar speeds, but I suspect software encoding, when slowed down, would still be superior. So if you want to encode videos, software encoding is still probably the best option in order to maximize quality and file size, whereas if you're talking about Twitch streaming, yeah, NVENC is probably better, or at least as good.
Posted by Mate
 - December 20, 2020, 18:26:17
@_MT_
QuoteEven if AMD was currently on the 5nm node, they wouldn't be able to match M1's efficiency.
Well, it depends how you define efficiency. According to TSMC 5nm improved energy efficiency by 30% and M1 is using 14W under load.  Ryzen Renoir(8 cores) can easily outperform Apple chip in general tasks with 20W so in theory it should be more efficient under load as 20*0,7 = 14.

However its clear that Apple aims to cover many use cases with dedicated coprocessors  that will make their processors very efficient. It is also probably easier to implement more coprocoessors for ARM  than in x86 as this makes x86 decoders even more complex. As things are now I can clearly see Macbooks in year or two just as better alternative for x86 premium laptops in performance/price ratio in every segment. AMD and Intel have very complex processor designs and it will be harder for them to keep up with  Apple each year.

@Vertigo
Maybe I used wrong words - I'm using Windows(for my personal computer) and MacOS side by side for years. However now M1 is very close to meet all my performance expectations(also for gaming!) and blow away x86 devices in terms of battery life and silence. My G14 would be already on ebay if m1 macbooks could  handle two external displays  in normal way. I'm really tired of that annoying noise and heating for no reason.
Posted by JayN
 - December 20, 2020, 17:54:33
There is a forbes article, "Intel's Surprise Ryzen Killer: Alder Lake Hybrid Processors Offer Biggest Performance Leap In 14 Years", in which Koduri gives his opinion on Alder Lake performance.

Jim Keller was excited about the wide out of order decode in Lakefield's Tremont.  Alder Lake is using Tremont's higher performance successor, Gracemont.  Perhaps it is continuing to build on the wide out of order decode feature.