Hacker News new | past | comments | ask | show | jobs | submit login

How many server models have HP and Dell switched to using Epyc?



HPE Solution Architect here -- HPE has 4 lines, 3 public / 1 private. DL385 and DL325 are rack based servers aka traditional pizza box.

DL385 is your workhorse platform. DL325 is a 1P design based for heavy PCIe connected (read: NVMe) devices.

The CL3150 is a cloudline server and will likely be more consumed in the service provider space, in my opinion.

The Apollo 35 is available to top 200 volume accounts (which has been annouced publicly, but is not available to your everyday customer.)

This type [of information leak] is my worst nightmare as someone who works with resellers, letting these types of documents in the wrong hands or somehow access is breached like this.

Edit 2: I am a server SA, MASE. I configure servers all the time. If customer demand shows the swap, you could see the proc move over to other lines, such as blades and the HPC markets.


Super-curious: have you seen much customer-driven demand for AMD EPYC, i.e. for reasons other than Xeon availability? AMD processors have been affected to a lesser degree to speculative execution exploits, they're cheaper and (if I'm not mistaken) offer more PCIe lanes, etc.

Also, do you expect 7mm EPYC to do well in your space? Thanks!


Depends on market segement. I've seen alot of demand for Epyc in workloads that are sensitive to memory bandwidth. Just so much to offer when you max out 8 channels.

They are cheaper due to fabrication process. The PCIe lane story is stuff of fanboys. It comes at a cost, power and heat.

Secondly, anyone looking at an NVMe box should be looking at AMD in my opinion. The trick is if you are doing a VM farm, mixing Intel and AMD aint the best idea, as you all know.

I see EPYC ticking up fast.

In terms of exploits like Spectre/Meltdown, I'm pretty sure the exploits AMD claimed were not vulnerable, they ended up pushing out microcode for anyway. So its a moot point.

I HAVE come across alot of customers who have DOUBLED their core count due to Spectre/Meltdown mitigations, and they are attracted to AMDs, high core, lower cost options. But remember, the power draw is different and always test/PoC!


> The PCIe lane story is stuff of fanboys. It comes at a cost, power and heat.

Could you unpack this a bit? Specifically, I'm curious if the cost is a premium per lane (e.g. W/lane greater on AMD than on Intel) [1]. Also, is that cost at all affected by the I/O volume or merely the CPU being power-hungry overall?

[1] Of course, that assumes everything else being equal, which it can't be, as well as equal proprotion of PCIe utilization, which is unlikely.


Ive had a few customers test AMD and found a higher operating temp and determined it was due to higher power consumption. On paper, you get more lanes at a lower TDP w/ AMD. In practice, as always, your results may vary. Test!

PCIe lanes and counting them is funny math. Do the homework on system boards, how they communicate, and the tax of moving information between processors.

However, I would say their tests were short, and AMD processors have 3 power operating modes. There was also a neat blog posted somewhere (I think on here...) a little while back suggesting that the AMD proc did not need to run at advertised power on the customer procs. It was about compile times and how much power still resulted in good times. That was consumer-grade Ryzen chips tested though.


Unfortunately, higher temperature says less about power and more about thermal design (often of the overall system and not just the chip).

> On paper, you get more lanes at a lower TDP w/ AMD.

I was hoping you (or anyone) had at least some real-world anecdata.

However, the theoretical power cost being lower suggests it's unlikely that if there's a premium in practice, it's unlikely to be significant.

> PCIe lanes and counting them is funny math. Do the homework on system boards

It's not that funny. Latency "taxes" are certainly a concern for some workloads, but, ultimately, if there's not enough bandwidth to get the data to the CPU, such that it might end up idle, that can trump any tax. The difference between 40 and 128 lanes of PCIe 3.0 in transferring 64MiB is on the order of 1ms.

Finding a mobo that allows access to all the lanes might be more challenging when there are 128 than when there are 40-48, but I expect the popularity of NVMe to reduce that challenge somewhat.

OTOH, it seems Epyc uses half those lanes for communication between CPUs, so the usable lanes doesn't go up for 2S vs 1S, so perhaps the comparison is really 128 lanes vs 96 lanes.


yes, latency vs. throughput, the main idea also behind GPU computing. It worked there well, and CPUs are incredibly going to sacrifice latency for throughout as well.


> PCIe lanes and counting them is funny math.

Do you have a "relevant" chunk of customers that are really looking for the high-density PCI-Express connectivity? Are the 128 lanes per system a feature that actually draws in users with real world demands or is this the wrong thing to focus on?

> There was also a neat blog posted somewhere

You must be talking about the DragonflyBSD mailing list: http://lists.dragonflybsd.org/pipermail/users/2018-September... (as linked by others by now).

To me this wasn't very surprising. It's well understood in the more technically inclined enthusiast community that underclocking Ryzen yields tremendous efficiency improvements. Famous overclocker "The Stilt" did a great analysis on Ryzen's launch day in 2017: https://forums.anandtech.com/threads/ryzen-strictly-technica... One of his benchmarks showed an almost 80% efficiency improvement when underclocking an R7 1800X to 3.3GHz, which is just above Epyc's maximum boost frequency. Since Epyc is almost the same silicon as Ryzen 1st Gen (B2 stepping instead of B1), the chips should have almost identical characteristics.

Unfortunately, I'm not aware of any similar detailed analyses on recent Intel Core processors to compare. Samsung's low-power manufacturing node used by AMD has often been cited as the specific reason for the steep efficiency curve (and the realtively low upper end compared to Intel), but the general trend is the same for almost all chips.

On the other end of the spectrum, overclocker der8auer measured about 500W draw in Cinebench when overclocking the Epyc 7601 to around 4GHz: https://redd.it/92u6db


> Are the 128 lanes per system a feature that actually draws in users with real world demands or is this the wrong thing to focus on?

I'm going to go out on a limb and suggest (based on my own experience[1]) that most users are too ignorant to know that this might be something that they want or would benefit from.

Some of us have always demanded more I/O bandwidth (even if it meant 4S servers), but typically with a price limit.

I do, however, suspect that additional demand could materialize in the form of NVMe slot count.

[1] particularly with so many potential employers being categorically cloud-only, they don't even want to know about the underlying hardware or what it's capable of.



> In terms of exploits like Spectre/Meltdown, I'm pretty sure the exploits AMD claimed were not vulnerable, they ended up pushing out microcode for anyway. So its a moot point.

But even in the scenario where the microcode actually did incorporate some "interesting" changes, they haven't impacted performance at all. So this is basically the world's biggest ever design win at this exact moment.


Thought I'd ask a followup question for the sake of others here who might appreciate the answer:

What sorts of evaluation opportunities do you provide for Intel vs AMD processor comparison?

Besides short-term on-site placement, this could even look like remote access to a specifically-provisioned lab environment.

Something like this would be pretty cool to evaluate a wide range of processor/memory/storage/etc combinations.

I'm absolutely sure this almost certainly exists but I thought it would be interesting know how it actually works. (I'm not in the server hardware industry, but am very interested in how it works)


Very nice. I will make sure our server teams know they can get Epyc. Hopefully we are already testing the Spectre impact on Xeon vx. Epyc.


I'd love to know your results privately if you would share, as I have several customers who were impacted by Spectre/Meltdown and are also doing private benchmarks.


I'm not super familiar with HPE's product lines but I'm just kind of surprised they are still using the Apollo name.

Maybe we'll see some Silicon Graphics branded HP gear too?


SGI stuff is already rebranded to the apollo name and Superdome Flex (UV300 I think it was before)

Depending on your SGI server, the name may have gone into the appropriate HPE family name.

If you need help with our new decoder ring, I'd be happy to point you in the right direction.

I expect HPC workloads to take advantage of AMD chips, but I do expect many of the mathematical improvements that the Skylake instruction set provides to amplify desire in the computational HPC/numerical computation/ non-GPU stuff.


Can you comment on the September/October availability claims?


I hate doing this, but because this is sensitive, I have copy pasta-d the HPE Customer statement from our Customer Presentation on the issue:

HPE, along with all other server vendors, is starting to see supply constraints on various Intel Xeon-SP processors—commonly referred to as Skylake—used in our Gen10 servers.

–Intel is supply tight on Skylake SP HCC, Skylake SP LCC, Skylake SP XCC, and Skylake W LCC processor series.

–Customer demand for various Skylake server processors is exceeding short-term supply.

–Intel is working diligently to increase supply but indicates limited ability to materially improve Skylake availability potentially through December 2018.


I'm not sure how many, but Dell has at least three in the PowerEdge line and they've been announced and available for a while. One of them is a 1U single-socket system with 32 cores, 64 threads, and up to 2TB of RAM.

https://www.amd.com/en/campaigns/amd-and-dell

https://www.dell.com/support/article/us/en/04/qna44314/dell-...

https://blog.dellemc.com/en-us/poweredge-servers-amd-epyc-pr...

http://www.itpro.co.uk/server-storage/30799/dell-emc-powered...


Dell seems less than enthusiastic about Epyc. For some reason their Epyc servers are 1/4th the density of their Intel servers. The dual socket Intel is available with 4 systems per 2U, the dual socket AMD is 2U.


They have long been an Intel allies since the days of Pentium Inside. Intel were paying billions to Dell to not have any AMD's product in their lineup. Dell even said the missing AMD were because there were no customer demands. Only after the AMD vs Intel case ruling, Dell decided to have one or two AMD machines just to make things look "better". My guess is in reality Intel runs deep inside their blood.


This may be attributable less to a lack of enthusiasm and more to the physical size and TDP of Epyc CPUs. Additionally, it would be very hard to take advantage of Epyc's increased PCIe lane count at half-U densities (which is, of course, not relevant to all use cases) and would likely constrain number of memory slots.

Personally, I've never been convinced that such high server densities [1] provide a net benefit. Ever since "blade" systems first came out, and through the current availability of the half-U models (e.g. 4 per 2U) or even 2S 1U models, they've been plagued with thermal design issues and, sometimes, reliability issues due to non-commodity parts and/or needing to spin tiny-diameter cooling fans so much faster.

Until relatively recently, there wasn't even datacenter space available that could accomodate such systems at full density running at full load. What did become available was at a premium (in addition to paying a premium for the higher-density hardware in the first place).

[1] more than 1S/U or so




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: