The AMD 7742 has 64 cores, and with certain models, they can be run in a "dual socket" server or workstation, meaning two AMD 7742 processors with 64 cores each on one motherboard - 128 cores in total. A "2x" in the description of a computer next to the procesor description usually signifies two processors on one motherboard.
Ryzen 3900X
Moderators: hgm, Rebel, chrisw
-
- Posts: 4889
- Joined: Thu Mar 09, 2006 6:34 am
- Location: Pen Argyl, Pennsylvania
Re: Ryzen 3900X
-
- Posts: 1080
- Joined: Fri Sep 16, 2016 6:55 pm
- Location: USA/Minnesota
- Full name: Leo Anger
Re: Ryzen 3900X
OK. Thanks.MikeB wrote: ↑Wed Sep 04, 2019 5:49 amThe AMD 7742 has 64 cores, and with certain models, they can be run in a "dual socket" server or workstation, meaning two AMD 7742 processors with 64 cores each on one motherboard - 128 cores in total. A "2x" in the description of a computer next to the procesor description usually signifies two processors on one motherboard.
Advanced Micro Devices fan.
-
- Posts: 1567
- Joined: Thu Jul 16, 2009 10:47 am
- Location: Almere, The Netherlands
Re: Ryzen 3900X
https://wccftech.com/intel-cascade-lake ... hedt-cpus/Zenmastur wrote: ↑Mon Sep 02, 2019 4:00 pmThat's not the point. The point is how much faster is the fastest car than a cheap car. In this case the difference in speed from a cheap car to the fastest car has been significantly reduced because the cheap car is now much faster.Joost Buijs wrote: ↑Mon Sep 02, 2019 7:20 amThe same reason why some people drive Lamborghini and others drive Austin cars.Zenmastur wrote: ↑Mon Sep 02, 2019 6:02 amIf you compare the dual Xeon Platinum 8280 system at 125M nps to the projected 106M nps of 3rd gen TR 32-core system one wonders why anyone would ever spend an extra $18,000 on CPU's for the Intel system.MikeB wrote: ↑Mon Sep 02, 2019 5:30 amIndeed,Dann Corbit wrote: ↑Fri Aug 30, 2019 2:46 am ...
And the bottom graphic here shows that the very expensive AMD 7742 smokes the AMD 7601 and utterly destroys the Intel 8280 in terms of nodes per dollar with SF and AF.
...
Regards,
Zenmastur
Regards,
Zenmastur
https://newsroom.intel.com/wp-content/u ... l-RWPE.pdf
-
- Posts: 201
- Joined: Thu Jun 06, 2019 8:05 pm
- Full name: Percival Tiglao
Re: Ryzen 3900X
Probably Databases.Zenmastur wrote: ↑Mon Sep 02, 2019 6:02 amIf you compare the dual Xeon Platinum 8280 system at 125M nps to the projected 106M nps of 3rd gen TR 32-core system one wonders why anyone would ever spend an extra $18,000 on CPU's for the Intel system.MikeB wrote: ↑Mon Sep 02, 2019 5:30 amIndeed,Dann Corbit wrote: ↑Fri Aug 30, 2019 2:46 am ...
And the bottom graphic here shows that the very expensive AMD 7742 smokes the AMD 7601 and utterly destroys the Intel 8280 in terms of nodes per dollar with SF and AF.
...
Regards,
Zenmastur
While EPYC and Threadripper are great for most workloads, consider how their L3 caches are split between "CCX" (core complex). You can only access 16MB of L3 cache per thread.
"External" L3 cache (between CCX) has been measured to be slower than accessing DDR4 RAM directly (!!). So any one thread can only access, at best, 16MB within 10-nanoseconds (the 16MB that fits in the "local" CCX L3 cache). In contrast, a single thread on a Xeon 8280 will be able to access all 38.5MB of L3 cache within 10-nanoseconds. In an L3 constrained environment (such as databases), this leads to better performance.
AMD EPYC / Threadripper can have 256MB of aggregate L3 cache, but its individual threads max out at 16MB of L3 cache. Now, this is a good thing for workloads such as Lazy SMP (only the "relevant" parts of the shared hash table get loaded into the local L3 caches of each thread). But its a bad thing for databases.
-
- Posts: 12549
- Joined: Wed Mar 08, 2006 8:57 pm
- Location: Redmond, WA USA
Re: Ryzen 3900X
https://www.businesswire.com/news/home/ ... ld-Records
"In a similar fashion for database virtualization, the HPE ProLiant DL325 delivered an astounding 321 percent performance boost over the previous record holder, providing a significant agility advantage in database provisioning and maintenance for cloud, big data and IoT environments."
"In a similar fashion for database virtualization, the HPE ProLiant DL325 delivered an astounding 321 percent performance boost over the previous record holder, providing a significant agility advantage in database provisioning and maintenance for cloud, big data and IoT environments."
Taking ideas is not a vice, it is a virtue. We have another word for this. It is called learning.
But sharing ideas is an even greater virtue. We have another word for this. It is called teaching.
But sharing ideas is an even greater virtue. We have another word for this. It is called teaching.
-
- Posts: 201
- Joined: Thu Jun 06, 2019 8:05 pm
- Full name: Percival Tiglao
Re: Ryzen 3900X
Database virtualization: yes. Because you split up your server into pieces and give it out to other customers. Which EPYC's "split" L3 cache design is outstanding, and its the ideal use.Dann Corbit wrote: ↑Thu Sep 05, 2019 10:06 pm https://www.businesswire.com/news/home/ ... ld-Records
"In a similar fashion for database virtualization, the HPE ProLiant DL325 delivered an astounding 321 percent performance boost over the previous record holder, providing a significant agility advantage in database provisioning and maintenance for cloud, big data and IoT environments."
However, if you run a bare-metal database, such as a single server with Redis, Memcached, or Postgres, the Xeon's "shared L3" cache becomes far more relevant. Larger, singular databases scale far better than trying to create replication clusters. Now granted: to "scale up" a server costs significantly more than to "scale out" with replecation / virtualization, but there's gains with regards to simplicity. Sometimes, spending $10,000 is cheaper than switching to a scale-out methodology (which may require programmers, aka $200,000+).
Its a niche, but bare-metal databases are still a thing. On the other hand, that 18-core / 72-thread POWER9 chip with 90MB L3 cache (!!) for around $2000 is better than a $10,000 Xeon with only 38.5MB of L3 cache. It seems like EPYC will win for cloud-instances and many multi-threaded programs, while POWER9 is superior for the memory-constrained / L3 constrained workloads.
Intel Xeon stays in the x86 ecosystem though, and still is better "scale up" than EPYC. I'm not saying AMD EPYC is bad or anything, I'm purposely focusing on a niche where an IT engineer might decide to go Xeon still. In most deployments, EPYC looks like a superior choice. But remember, EPYC is a "Scale out" platform. Split L3 caches, CCX, etc. etc. Its great for the modern containerized / virtualized workload, but not everyone has switched to containers / scale out designs yet. (And besides: scale up is simpler to design and run: fewer servers often means less downtime)
-
- Posts: 12549
- Joined: Wed Mar 08, 2006 8:57 pm
- Location: Redmond, WA USA
Re: Ryzen 3900X
Taking ideas is not a vice, it is a virtue. We have another word for this. It is called learning.
But sharing ideas is an even greater virtue. We have another word for this. It is called teaching.
But sharing ideas is an even greater virtue. We have another word for this. It is called teaching.
-
- Posts: 919
- Joined: Sat May 31, 2014 8:28 am
Re: Ryzen 3900X
At first I thought you were just trolling. But...dragontamer5788 wrote: ↑Thu Sep 05, 2019 5:37 pmProbably Databases.Zenmastur wrote: ↑Mon Sep 02, 2019 6:02 amIf you compare the dual Xeon Platinum 8280 system at 125M nps to the projected 106M nps of 3rd gen TR 32-core system one wonders why anyone would ever spend an extra $18,000 on CPU's for the Intel system.MikeB wrote: ↑Mon Sep 02, 2019 5:30 amIndeed,Dann Corbit wrote: ↑Fri Aug 30, 2019 2:46 am ...
And the bottom graphic here shows that the very expensive AMD 7742 smokes the AMD 7601 and utterly destroys the Intel 8280 in terms of nodes per dollar with SF and AF.
...
Regards,
Zenmastur
While EPYC and Threadripper are great for most workloads, consider how their L3 caches are split between "CCX" (core complex). You can only access 16MB of L3 cache per thread.
"External" L3 cache (between CCX) has been measured to be slower than accessing DDR4 RAM directly (!!). So any one thread can only access, at best, 16MB within 10-nanoseconds (the 16MB that fits in the "local" CCX L3 cache). In contrast, a single thread on a Xeon 8280 will be able to access all 38.5MB of L3 cache within 10-nanoseconds. In an L3 constrained environment (such as databases), this leads to better performance.
AMD EPYC / Threadripper can have 256MB of aggregate L3 cache, but its individual threads max out at 16MB of L3 cache. Now, this is a good thing for workloads such as Lazy SMP (only the "relevant" parts of the shared hash table get loaded into the local L3 caches of each thread). But its a bad thing for databases.
In the end the only measure that makes a difference is how it performs on your specific application. I don't care if it takes an hour per cache transfer as long as my programs complete their workload in less time. In most cases, EPYC seems to do exactly that AND at less than half the cost. You can play with the numbers all you like, the only thing that really matters is does it get the job done cheaper and/or faster. For small business and cloud customers it seems like it's by far the best choice for most applications.
Regards,
Zenmastur
Only 2 defining forces have ever offered to die for you.....Jesus Christ and the American Soldier. One died for your soul, the other for your freedom.
-
- Posts: 201
- Joined: Thu Jun 06, 2019 8:05 pm
- Full name: Percival Tiglao
Re: Ryzen 3900X
Hmm, I was going off of pgbench results from last gen and my expectations for Zen2. But if TPC has a set of results for the new generation, that overrides what I was going off of. Good to know.
I wish there was a 1x Xeon result. There's only a 2x Xeon result: http://www.tpc.org/tpch/results/tpch_re ... =117111701
2x28 Core setup outperforms the 1x32-core Zen2 result: http://www.tpc.org/tpch/results/tpch_re ... =119080501
There's not an apples-to-apples comparison, unfortunately. The other Zen2 results for 3000 GB are for "EXASOL" Database, not SQL Server 2017. So the above two results seem like the best apples-to-apples comparison. But 2x 28Core vs 1x32 core leaves a lot to be desired...
http://www.tpc.org/tpch/results/tpch_pr ... lttype=all
----------
EDIT: Also, those benchmarks are "bigger" than I was expecting. I was thinking more about RAM-sized databases. Once the dataset grows beyond the size of RAM, EPYC gains an advantage due to having more PCIe lanes than Xeon. Hmm, now that I think of it, EPYC seems to have a lot of advantages.
-
- Posts: 1567
- Joined: Thu Jul 16, 2009 10:47 am
- Location: Almere, The Netherlands
Re: Ryzen 3900X
I wonder when Zen 2 will be readily available, 3900X is almost impossible to get, 3950x is still not announced, and the processors that are released have trouble to reach the specified turbo speed. As ususal it is drama with AMD, I'm not an Intel fan whatsoever, but if Intel delivers 'Cascade Lake X' for a price that compares favorably to AMD Zen 2, I will go that way.