Review Index:

AMD Trinity Review: The A10 5800K and A8 5600K Break Cover

Author: Josh Walrath
Subject: Processors
Manufacturer: AMD

Results: 7-Zip and Valve Benchmarks



This open source file extractor and compressor has a very handy benchmark to show the relative MIPS of differing processors and thread counts.  This program is updated quite often to address changes in processor architectures.

View Full Size

All of the AMD parts are very, very close together throughout testing.  Only when we get to four threads do we see any separation.  The four “real” cores of the 3870K compare very well to the 2 module design of Trinity.  We also see very similar performance between the 5800K and the 4170.


Valve Source Particle Simulation

Valve created a couple of benchmarks which can leverage CPUs with multiple cores.  The first of which is a particle simulation run within the Source engine.

View Full Size

The more cores the merrier.  The i3 2105 also has a very strong showing as compared to the AMD parts, except for the six core 1090T.  Note again the lack of separation with the FX-4170 and 5800K.


Valve Source Map Compilation

The second test compiles a map using multiple cores.  This applies static lighting and other features throughout the map so it looks pretty once on the screen.

View Full Size

The X4 980 is significantly faster than the rest of the field with the 1090T coming up right behind, and perhaps most surprising the 3870K.  The 5800K is slightly faster than the 4170, but both are much slower than the i3 2105.

October 2, 2012 | 12:48 AM - Posted by tbone (not verified)

did you turn off turbo core in bios?

also I'm sure a bios update is on the way

also make sure you use a good heatsink cause it will throttle

October 2, 2012 | 01:00 AM - Posted by Anonymous (not verified)

Yea, Tigerdirect has them

October 2, 2012 | 01:08 AM - Posted by rahul garg (not verified)

I think the units for SiSoft Sandra memory bandwidth results should be GB/s, not MB/s.

October 2, 2012 | 01:51 AM - Posted by Arb (not verified)

so what was the power usage when OC'ed to 4.4ghz ?

October 2, 2012 | 12:57 PM - Posted by Josh Walrath

I will be doing more testing throughout the week of overclocked performance.  Apparently there are a few settings which are deletrious to good overclocked results that we were unaware of (like the above mentioned Turbo Core).  Apparently there are others as well.  Once I get a better lock on overclocking, I will throw up some numbers here.

October 2, 2012 | 03:56 AM - Posted by pat (not verified)

They need to drop these into some x86 tablets. That's where the real advantage comes I think.

They also need to do a deal and get these in tv or tv boxes.

October 2, 2012 | 07:20 PM - Posted by aparsh335i (not verified)


October 2, 2012 | 04:00 AM - Posted by Nilbog

Great review, thanks again Josh.
I was afraid this would happen again. I really want AMD to pull ahead or at least catch up.
Barely keeping up with an i3 isn't keeping my hopes up. I would buy a new AMD CPU in a heartbeat if it could even come close to an i7. I want AMD to win dammit.
I realize that they are constantly working on the next CPU, i really hope somewhere down the line they just stop working on this architecture and go back the the Phenom II architecture (K10 I think) or at least re-release it along side bulldozer.
Is that possible? Or just too costly?
To this day they are still faster, just more power hungry. Not much of a down side IMO.
Why aren't they pumping more cores in stuff? That would help a little wont it?

EDIT: oooh didnt see the tablet comment, that would be a great idea. Trinity Win8 tablets would be a huge advantage for AMD. I'm sure they can keep up with an Atom.

October 2, 2012 | 07:25 PM - Posted by aparsh335i (not verified)

I don't want to rain on your parade, but i really have to disagree with a few things you said. Going back to the phenom II architecture = throwing in the towel and going out of business. AMD is NOT faster and more power hungry. It's slower and more power hungry. Why aren't they pumping more cores? They did do that with Bulldozer, and it has not been what people want as far as performance goes for MOST (some may still be appreciating 8 cores for small niche groups).

October 2, 2012 | 09:17 PM - Posted by Nilbog

You misunderstood what i meant. Sorry, to be clear. I was talking about the Phenom II architecture being faster and more power hungry, in relation to AMDs new CPUs (it still pulls far ahead in most of the benchmarks).

Its quite obvious they are slower and less efficient than Intels offerings.

I bring up going back to the Phenom II architecture so that AMD can actually compete. Instead of being the "Affordable" CPU manufacturer, and not competing at all.

I was thinking more along the lines of 10 cores minimum (Clearly 8 wasn't enough) or perhaps im being naive thinking that you can just throw more power at the problem.
For the record i wold be SO SOLD on a 10 or more core consumer part. Its just cool IMO, fastest or not.

October 4, 2012 | 03:45 PM - Posted by Lord Binky (not verified)

This is my belief based off of their releases so far.

They knew where they were at in generalized ‘speed’ or processing power, and new their architecture could not continue scaling into higher clocks and higher core counts and keep up with hyperthreading. I really think AMD tried to make an architecture that scaled well for multiple cores and high clock speeds. I think their processor lines show they succeeded in this, it is almost arbitrary for them to include more cores.

Now they are optimizing that core architecture (increased Instructions Per Clock, etc.) and that is the effort that will bring them back into the performance game. Running more real cores, higher overclocks (5Ghz+ wooo!), and close to or better IPC, will make people wonder WTF Intel has been doing. They have licenced Intel’s 3D trigate patents so they can continue scaling down with their manufacturing processes too.

AMD chose the less risky route here by optimizing their core count scalability then optimizing core operation, but it is also the longer slow & steady route. If they had tried to create Optimized high clock high IPC cores it is MUCH riskier.

They would either be back to dual core counts with the possibility their design does not scale into high core counts in which case they have to scrap that work when trying to design a high core count processor.

Or they are try to design an archecture that does it all and the results of that all coming together for one processor design is very risky in terms of ready time and design performance, which has a high possibility to leave them with a mediocre architecture they are unable to build upon to continue competeing.

October 2, 2012 | 05:45 AM - Posted by dragosmp (not verified)

Just to complete the picture, it is rumored that AMD tried to solve the major bottlenecks of the Bulldozer with Piledriver: scheduling and L2 cache speed. There are some tests at xbitlabs AIDA64 memory benchmark - the results show that in most Bulldozer and Piledriver are similar, but L2 read bandwidth of Piledriver has risen dramatically, almost doubled:

This could be a major reason (along with double the scheduling queue) why the A10 that has no L3 cache managed to stick quite close to the FX 4170.

Anyway, this probably doesn't bring that much of a performance benefit to make Piledriver look good, but as far as I remember this is the first time since Clawhammer (anyone remembers that one?) that AMD hasn't degraded the speed or latency of the cache on a next-gen CPU. Let's hope the L3 in Vishera continues on this route.

October 2, 2012 | 06:32 AM - Posted by Prodeous (not verified)

Another great review from a great man.. Well close enough :P

Still as the review goes, I do wonder how much lack of L3 cache has on performance. I know AM3+ Piledriver is still a few weeks away, I just hope it will make an upgrade from my 1090T worth it.

Still any speculations how much L3 cache could improve in terms of CPU performance? 5-10%?

Is there a way to test Buldozer cores with L3 disabled and enabled to see how much performance hit there was?

Now back to Trinity. New socket is a killer for me, but it is nice that AMD accnowledged this and stated that there will be at least one more CPU architecture (steamroler?) for the socket before AM3+ and FM2 fuse into one socket.

If only Blender3D would support OpenCL rendeirng on AMD GPUs, I would be inclined to think of this setup. For now only mobile Trinity seams to make sence at this point.

October 2, 2012 | 11:32 AM - Posted by HUU (not verified)

the results of you cinebench r10 for the i3 2105 are MUCH, MUCH LOWER than the normal score...

please re run the benchmark, make sure to test it in 64bits.

October 2, 2012 | 02:48 PM - Posted by Ryan Shrout

We are redoing tests here and will look into it.


October 2, 2012 | 03:57 PM - Posted by Anonymous (not verified)

This is a very narrow product line - nothing more than quad-core? The price varying only from $71 to $122? No part less than 65W? Aimed almost only at gamers.

Obviously the objective here was to get the FM2 socket exactly right so it can remain stable until late 2014. By that time, Thunderbolt has either replaced all these various nasty display ports, or it has not. Also by that time, PCIe 3.0 x32 devices will be more common and the full 52GB/s of the HTX bus should be available at the DDR3, DDR4 or DDR5 RAM interface. And, 10 gigabit ethernet will either have come down in price (thanks to Thunderbolt) or be irrelevant to desktops and NAS (thanks to Thunderbolt). By 2015, with these existing buses all maxed out, and the IEEE P1905.1 standard settled so that things like powerline networking's interface to the PSU can be settled, and Thunderbolt vs. DisplayPort vs. 10 gigabit Ethernet settled somehow... FM3 or whatever can be stable from 2015 to 2018.
Maybe longer. The desktop will be dead by then, and the mainboard is going to be sitting in your wall near your electrical box or cable head, and talking to your refrigerator as much as to your TV.

Expect some ARM cores (for very low power idling) in the very next FM2 processor release. That's the only way to respond to Haswell. Expect also some X6 X8 X10 and X12 processors in that lineup, and a few low-power options below 50W (with the ability to rely on ARM core to respond to routine network and device events to keep that power draw much lower in practice). Much more price variations, perhaps from $50 to $200 or even $300.

Given the graphics performance of these October 2012 chips though, it's entirely reasonable to rely on the embedded graphics and use the PCIe x16 slot for a PCIe SSD - basically equivalent to slower RAM given the FM2 direct chip connection. Imagine 100GB RAM or 250GB RAM for a few hundred bucks (some OCZ PCIe SSDs sell for as little as $2/GB so that's $200-$500, same as a good video card).

Given the excellent multi-core performance of database engines, and the very low price of these chips, it's possible you could see lots of FM2 processors used in database hosts. Especially if there is a way to use OpenCL to do the processing on all those shader cores...?

October 2, 2012 | 06:43 PM - Posted by allpeoplesuck (not verified)

This is a good preview for Vishera Piledriver. When that time comes, can see see the desktop Piledriver Vishera review with a Core 2 Duo (like Q6600), i7 920, 2600k, 3770k, and Phenom 2 quad and hex (and of course old FX).

October 2, 2012 | 11:17 PM - Posted by Anonymous (not verified)

AMD's Trinity platform is a good platform, yes it trades blows with the i3 with Intel's chips hitting hard when it comes to single threaded applications. However AMD hits Intel hard on entry level gaming.

When it comes down to Power Consumption I feel that the whole story isn't being published. Intel's HD 4000 just doesn't cut the mustard to games and basically requires a ext. video card to edge out the AMD APU. With that said I have yet to see a power consumption table to show what the i3 or i5 have with an ext. video card. AMD's APU already has a full blown video card on die and reflex it in it's power consumption. Intel's on die GPU is to show lower power consumption on the charts but knowing full well no one in there right mind would would run it that way.

October 3, 2012 | 04:34 AM - Posted by Arb (not verified)

true sad thing of these tests, fact that no one in their right mind uses intergraded cpu graphic's to play games. Yet AMD seems happy to beat intel in these, since its only thing they can win in.

October 3, 2012 | 11:23 AM - Posted by Josh Walrath

All kinds of people in their right mind play games with integrated graphics.  My children can play all their DC Universe, Roblox, Hero-Up, and all kinds of other 3D games which run perfectly fine on integrated graphics.  They don't play Crysis, or BF3, or Skyrim, but the games they play are designed from the ground up to be played comfortably on integrated grahpics.  I'm actually impressed by how DC Universe looks on an APU.  Plays pretty smooth, looks good, and it is a cheap platform for users.

October 3, 2012 | 10:21 PM - Posted by tbone (not verified)

keep us updated on the overclock hunt

have you tried a better cooler?

cpu OC? igpu OC? memory OC?

how does it perform when overclocked?

be cool to do a video review of the overclock ;)

October 6, 2012 | 01:36 AM - Posted by VikingFteo (not verified)

Looks like Llano 3870K is a much better buy at this stage. One can just clock its gpu to 900Mhz and blow away all the Trinity A10-5800K benchmarks above!. Llano does not seem to perform as well in the CPU side when clocked to 3.6Ghz from 3Ghz. The fact the Trinity is so highly clocked means that the K parts are pretty useless as the OC headroom is very small. Shown here only 4.4Ghz max which is really poor. This shows that AMD is putting a small headroom "buffer" in their Trinity chips. Intel , however, can be clocked to such frequencies fairly easily!.

October 13, 2012 | 03:51 AM - Posted by Anonymous (not verified)


I own an AMD A8 5600k and i want to ask you what video card to choose from these: varianta dual fan

I want to know if i can use all of them in dual graphics mod, AMD recomends HD6570 and HD6670 for this processor


September 24, 2014 | 02:56 PM - Posted by Tammi (not verified)

Very great post. I simply stumbled upon your blog and wished to mention that I've really enjoyed surfing around your blog posts.
After all I'll be subscribing in your rss feed and I'm hoping you write again very soon!

Stop by my web page :: instagram