Faster, probably, but 2x+ Faster like in PowerPC > Intel very unlikely. Which is why there’s much less likely to be enough performance budget for effective emulation.
If they could just turn up the TDP and clock speed on their chips and get 2x the performance of Intel's best chips that easily- they would already own the entire desktop market.
As someone without a CPU design backround, would something like this scale pretty linearly with added power and thermal headroom? I assume there are limit that would have to be overcome, but what would an ARM chip in the conditions of an i7 look like?
TDP typically scales as somewhere between the cube and fourth power of the clock speed if you're pushing the envelope (in the sense of running at frequencies where further frequency increase also needs a voltage increase). So having 10x the thermal envelope means you can probably clock about twice as fast, all else being equal.
Computer architectures routinely see 3x performance jumps across different power budgets. This rule has held over decades.
Clock speeds alone can probably increase by 30%. Caches and internal datapaths can double or more. Then you can start to add in more execution units or more expensive branch-prediciton or even new more power-hungry instructions.
A 4 Watt Intel Pentium 4410Y Kaby Lake for mobile devices gets about 1800 on Geekbench, while a 115 Watt Intel Core i7-7700K Kaby Lake for desktops gets 5600.
I'm just going to say it: the Apple laptop CPU is going to get Geekbench score... above 9000!
So artificial benchmarks already do a very poor job of capturing performance. The apple laptop cpu does not exist. If it did exist it would likely suffer a very substantial performance hit if forced to emulate x86 software. So you are going to speculate on the meaningless benchmark numbers of an imaginary cpu that will take a wholly unknown hit if everyone does not rewrite everything why?
I suspect if Apple designs a desktop CPU, performant x86 emulation will be a key design criteria. I know very little about CPU design, but I imagine it would be possible to have hardware optimisations for x86 emulation just like we have today for video codecs.
Or even further they could bake a "rosetta" into the chip's microcode and have their CPU natively support the x86 instruction set along with ARM or whatever they come up with.