Review Index:

Frame Rating: Eyefinity vs Surround in Single and Multi-GPU Configurations

Manufacturer: Various

Summary of Events

In January of 2013 I revealed a new testing methodology for graphics cards that I dubbed Frame Rating.  At the time I was only able to talk about the process, using capture hardware to record the output directly from the DVI connections on graphics cards, but over the course of a few months started to release data and information using this technology.  I followed up the story in January with a collection of videos that displayed some of the capture video and what kind of performance issues and anomalies we were able to easily find. 

My first full test results were published in February to quite a bit of stir and then finally in late March released Frame Rating Dissected: Full Details on Capture-based Graphics Performance Testing which dramatically changed the way graphics cards and gaming performance was discussed and evaluated forever. 

Our testing proved that AMD CrossFire was not improving gaming experiences in the same way that NVIDIA SLI was.  Also, we showed that other testing tools like FRAPS were inadequate in showcasing this problem.  If you are at all unfamiliar with this testing process or the results it showed, please check out the Frame Rating Dissected story above.

At the time, we tested 5760x1080 resolution using AMD Eyefinity and NVIDIA Surround but found there were too many issues and problems with our scripts and the results they were presenting to give reasonably assured performance metrics.  Running AMD + Eyefinity was obviously causing some problems but I wasn’t quite able to pinpoint what they were and how severe it might have been.  Instead I posted graphs like this:

View Full Size

We were able to show NVIDIA GTX 680 performance and scaling in SLI at 5760x1080 but we only were giving results for the Radeon HD 7970 GHz Edition in a single GPU configuration.


Since those stories were released, AMD has been very active.  At first they were hesitant to believe our results and called into question our processes and the ability for gamers to really see the frame rate issues we were describing.  However, after months of work and pressure from quite a few press outlets, AMD released a 13.8 beta driver that offered a Frame Pacing option in the 3D controls that enables the ability to evenly space out frames in multi-GPU configurations producing a smoother gaming experience.

View Full Size

The results were great!  The new AMD driver produced very consistent frame times and put CrossFire on a similar playing field to NVIDIA’s SLI technology.  There were limitation though: the driver only fixed DX10/11 games and only addressed resolutions of 2560x1440 and below.

But the story won’t end there.  CrossFire and Eyefinity are still very important in a lot of gamers minds and with the constant price drops in 1920x1080 panels, more and more gamers are taking (or thinking of taking) the plunge to the world of Eyefinity and Surround.  As it turns out though, there are some more problems and complications with Eyefinity and high-resolution gaming (multi-head 4K) that are cropping up and deserve discussion.

Continue reading our investigation into AMD Eyefinity and NVIDIA Surround with multi-GPU solutions!!

September 17, 2013 | 07:22 PM - Posted by Alien0227, I am one of those "that are invested in the HD 7000-series of graphics cards and have already purchased their Eyefinity configurations on the promise of a great experience".

I thank you for your hard work PcPer! I am truly grateful!

Because of it so far, I have seen a dramatic improvement in single screen Xfire game play with my 2 HD 7870s.

I did however invest in triple 1920 x 1080p screens, 2 HD 7870s and I wish I could be rewarded with the performance expectations I paid for. The experience would be awesome.

I only hope AMD will strive to keep me as a loyal customer by listening to your findings and offer a solution shortly to this Eyefinity-CrossFire problem.

You and AMD will have my undying loyalty and gratitude for it.

If AMD does, it will surely make them the king of the hill in value, for dollar per dollar performance...

Are you listening AMD? Is this make or break time for your company? My next GPU purchase depends on how you react. I wish AMD great success!

p.s. Ryan and Team. Please be confident that you are doing a great service to AMD in the long run and their customers.

Marc Senecal
Fort Myers, FL

September 17, 2013 | 09:42 PM - Posted by Ryan Shrout

Marc, thank you very much for your comments and feedback.  We greatly appreciate the support!

September 18, 2013 | 04:10 AM - Posted by j (not verified)

I'm sorry but I have to correct you: gaming equipment is not an INVESTMENT. Unless you're a pro gamer is a consumer spending.

Anyway, I wish you all the best fun with your new hardware :)

September 18, 2013 | 10:21 AM - Posted by Aranyszin (not verified)

This is no place for a grammar Nazi, and you aren't even correct in the first place.

Yes, the word "investment" is often associated with money spent expecting a monetary return. However, that is not the only usage of the word.

The poster clearly is "invested" considering the amount of money he spent, which he spent anticipating a return - in this case, a return in performance.

Honestly, contribute something useful to the topic or stay quiet. This thread is about obtaining a correction to a product flaw, not "How to parse nomenclature".

September 18, 2013 | 10:03 PM - Posted by Roy Taylor (not verified)

Dear Marc, yes we are definitely listening. Drop me a line direct.

best regards

January 28, 2014 | 08:20 AM - Posted by Roy Taylor (not verified)

hi Marc, I can assure you that myself and several other senior executives and engineers at AMD are deeply committed to making sure you get the experience you deserve.

We believe that Ryan treats this subject fairly, reports accurately and he works closely with us. We are grateful of his recognition of our leadership where we have it, even if I dont personally agree about our competitors 'sexyness'!

We will back with more on this shortly,

September 17, 2013 | 09:33 PM - Posted by Jonny (not verified)

there is 13.10 beta did u test it?

September 17, 2013 | 09:42 PM - Posted by Ryan Shrout

Nope, we were using 13.8 beta.  But 13.10 did not add any changes for Eyefinity / Frame Pacing.

September 18, 2013 | 01:43 PM - Posted by Jonny (not verified)

i got 7950 X2
I had a chance to buy another one at $ 100

Now I'm all out of it and save the next generation of NVIDIA ...
I'm very disappointed that AMD are selling a defective product ..

Thanks for the information.

September 18, 2013 | 04:44 PM - Posted by NvidiaPWNS (not verified)

AMD is always all about selling defective and inferior products. They did it before with their "Tri core" CPUs. Those were just Quad cores that had a 4th core that they couldn't get to work. AMD GPUs and drivers are shit. You get what you pay for.

September 19, 2013 | 06:05 AM - Posted by Anonymous (not verified)

You do realize that binning is a key part in the silicon-industry? By your logic, the Titan is crap because it only has 14 of the GK110's 15 SMX units activated.
When a produc-line is announced, there are actually few different dies being produced. 3820, 3930, 60 and 70 are all 8-core "xeons" with cores disabled due yield issues.
EVERYONE do this.

And please, stop spreading the false claim that the drivers are bad. Maybe they were back in good ol' '04, but that is long gone. AMD has actually had better drivers than Nvidia this generation...

September 21, 2013 | 11:13 AM - Posted by NvidiaPWNS (not verified)

In your dreams retard! NVIDIA PWNS YOUR AMD, NUB!

September 18, 2013 | 05:22 PM - Posted by mAxius

Are you sure

I'm still wondering about this little gem... "PCI-E bus speed is no longer set to x1 on the secondary GPU when running in CrossFire configurations"

September 19, 2013 | 12:24 AM - Posted by JJ (not verified)

"But 13.10 did not add any changes for Eyefinity / Frame Pacing." - I am curious about this too, as the press release clearly states that it updated something regarding Eyefinity.

September 19, 2013 | 12:50 AM - Posted by JJ (not verified)

I'm fairly technical, but I am getting a little outside my level of knowledge here

Each lane of PCIe 3.0 only has 1 GB/sec

This is a very old article, but THG did testing graphics cards by limiting the PCIe lanes available to graphics cards, and you can see a very large performance degradation.

3. Back to the "PCI-E bus speed is no longer set to x1 on the secondary GPU when running in CrossFire configurations" Note in the press release.

Would one of the cards in the setup being starved for bandwidth be able to account for these anomalies?

September 19, 2013 | 01:35 AM - Posted by Ryan Shrout

By AMD's admission, no. This problem listed in the Catalyst 13.10 notes only affects Crossfire configurations that do not use a bridge adapter across the graphics cards. This coming from AMD's own Andrew Dodd.

If this were a fix for our problems AMD would surely be crowing about it.

September 17, 2013 | 10:53 PM - Posted by Anonymous (not verified)

You guys rock pcper. Been here for years. Love amd, but they need to get their act together. Emphasis on the good faith part ya know. Don't become the crappy option. Amd has so much great iP if they could only get their software side together they would be SIGNIFCANTLY more competitive.

September 17, 2013 | 11:05 PM - Posted by Nacelle

I hope this fix AMD says they are working on will help with my 6970's too. I know they aren't worth that now, but I paid $700 for the pair, a couple years ago, to power my 3 screens. I'm about to sell them on craigslist and get an nvidia card, if it doesn't.

September 18, 2013 | 12:10 PM - Posted by Ryan Shrout

I honestly don't know if they will fix it, but the 13.8 beta did fix single screen for 6000-series parts so you might get lucky here too.

September 17, 2013 | 11:10 PM - Posted by Paul Keeble (not verified)

Keep hitting them until they fix this. I bought 2x 7970's in January 2012 and I noticed immediately the problem. Its not like AMD has only known about this since the beginning of this year, thousands of users were reporting this problem a year before that. It really put me off dual cards until I got a pair of 680's and found that the grass was much greener on the Nvidia side with SLI.

We need this fixed so we have some competition in the high end of GPUs.

September 18, 2013 | 08:40 AM - Posted by Anonymous (not verified)

Ryan...let us see your Tiger-Bus !

September 18, 2013 | 12:13 PM - Posted by Ryan Shrout

How about this?


September 18, 2013 | 09:51 AM - Posted by Anonymous (not verified)

with any luck a fix will be out from AMD a few after the release of the 9000 series, they are so slow with driver updates that I'd see it taking that long ... if they dont just give up ...

September 18, 2013 | 09:52 AM - Posted by Anonymous (not verified)

a few *months* after

January 17, 2015 | 07:33 AM - Posted by Merlin (not verified)

Many of the homebuyers tend to miscalculate the amount of money they borrow, most of the time, overestimating
them, because they think that their income will increase after several years and
that will make the mortgage payment be more comfortable
for them as the time goes by. Qualified properties are identified on the Help-U-Sell website, . Each house loan program will have its own individual set of rules. The upsells are also far cheaper than the actual cost of the phone. One of the nice things about the Federal Housing Administration loan, the FHA loan, thats the first time home buyer type loan, the minimum down payment loan, its only 3 years after you have had a foreclosure that you can qualify to purchase a home again. To begin with, they grant flexible finance guidelines which permit buyers having low credit history to get approved.

Here is my website ... como perder barriga rapido

September 18, 2013 | 10:23 AM - Posted by NLPsajeeth

So you can't run 3 PQ321's on AMD in a 6x1 eyefinity configuration?

I'm glad to see AMD is putting focus on 4K and hope they have a 4K eyefinity solution soon.

The only way NVIDIA will ever support anything other than 3x1 surround is if AMD turns up the heat. NVIDIA if you are listening, you need 2x1 and 2x2 surround support at any resolution to stay competitive on the consumer side. No one is dropping thousands of dollars on Quadro's just to get that one feature.

September 18, 2013 | 12:15 PM - Posted by Ryan Shrout

I think technically YES you can support 3 4K monitors like the ASUS PQ321Q with AMD today...but the main issue is going to be PERFORMANCE.  Even without frame pacing and interleaving problems how much horsepower would you need for that?

More than they have available today.

September 18, 2013 | 02:50 PM - Posted by NLPsajeeth

Just wanted to make sure it was a performance issue not a driver issue. I agree you wouldn't be able to run anything more than simple 3D demos in such a setup.

Agreed, I really hope we will see large GPU performance increases with the next round of silicon so that multi 4K gaming becomes a reality.

Keep up the good work, you guys are really pushing the boundaries of what is possible with 4K!

September 18, 2013 | 10:30 AM - Posted by gamerk2 (not verified)

I can't help but wonder, if the issues seen really are due to lack of proper synchronization, then would there be a FPS impact when AMD makes their changes?

Or in other words: Is AMD cheating on performance (knowingly or not)?

September 18, 2013 | 12:16 PM - Posted by Ryan Shrout

I don't think they were doing it knowingly, no.  That doesn't excuse them though; they should have been able to see these problems before and been trying to fix them before this all started to happen this year.

September 18, 2013 | 12:05 PM - Posted by fhmuffman (not verified)

Hi Ryan, I hope 4K connectivity is scheduled to be included in all future reviews of hardware, like laptop reviews. Back in June I needed to buy something quick when my system crashed and it would have been great to know if any low to mid range laptops could at least drive a 4K display. I am not expecting benchmarks of a Chromebook running Metro Last Light at 4K but it would be nice to know if I could display Sheets on a Seiki display with an A10 based laptop.

September 18, 2013 | 12:18 PM - Posted by Ryan Shrout

The SEIKI's should be supported with just about any modern GPU (single head, 30 Hz) but it's the 60 Hz models that require some more testing.  I'll try my best to include them going forward!

September 18, 2013 | 02:09 PM - Posted by Batman (not verified)

Thanks Ryan & PcPer for doing this & the previous investigative work; it is much appreciated & good to see AMD taking the findings onboard to make fixes!

September 18, 2013 | 08:25 PM - Posted by Pete (not verified)

If you used a monitor that was capable of 600fps would the problem persist?

September 18, 2013 | 09:31 PM - Posted by Ryan Shrout

Yes, the supported refresh rate is irrelevant here.

September 18, 2013 | 08:42 PM - Posted by Serpent of Darkness (not verified)

@ Ryan Shrout,

1. You need to state if your using AMD Beta Driver 13.8A (8-1-2013), or AMD Beta Driver 13.8B(8-19-2013). If you're using 13.8A on purpose during a discussion/benchmark on Surround and 1600p, multiple viewers could come to the conclusion you did this on purpose to make AMD look bad. AMD Beta Driver 13.8A doesn't have 1600p support. It only addresses the issues for DX11 API. 13.8B addresses 1600p and surround, if I am not mistaken. A possible upcoming 13.8C may address DX9 API, or it could have already been done in the new 13.9 WHQL update.

2. Personally, I can't take your discussions on a more serious manner. In your conclusions, you state things that give me the impression that you don't fully understand graphs, or have poor views of AMD Graphic Cards. At the very least, it is leading me to believe that you are bias towards Nvidia. Having favoritism, or a bias point of view to one company over the other, isn't a good way to approach a discussion or benchmark on any product. It doesn't help you seem serious, experienced, or reasonable to both bases (AMD and Nvidia users). It only tells readers that you pander to one side, and talk crap about the other brand's short-comings. AnAndtech doesn't do it, Guru3D doesn't do it, doesn't do it either, and they all come out with really good benchmarks about computer-based products. Both bases read their benchmarks because they aren't bias. Mr Shrout, you are bias either because you are letting people know of your hatred towards AMD, or you want to cater discussions and benchmarks that make AMD look bad to the Nvidia Base. Those are reasonable conclusions. If I don't see a benchmark on here discussing why the GTX 600, 700 and Titan series doesn't fully support DX11.1(support only software, but not hardware-wise), you are only going to prove me right.

3. Looking at the Frametime Variance Graphs that you posted, AMD 7970 will have a lower minimum band because the cards push lower latency to produce batches of frames. Problem with it, and it's true, is somewhere along the way, they will produce "runt frames." Frames that aren't one whole frame. It could be like 0.8 frames, or 0.9, or 0.7. On the other hand, it takes less time for AMD video cards to produce those batches of frames. Nvidia takes longer to produce the batch because, hardware wise, the system probably calculates whether it needs to spend more time producing an extra "whole" frame. That's why their minimum frame time band is higher than AMD. The hardware is always trying to push 1.0 frames times x amount of frames to a batch.

September 18, 2013 | 09:30 PM - Posted by Ryan Shrout

1. You are incorrect in this assumption.  No beta or full release of driver from AMD addresses Eyefinity.

2. I don't understand the relevance to DX11.1 reference here honestly.  This story isn't about API support but rather multi-display + multi-GPU gaming.  As to the bias question, this is something that gets targeted at people all the time when their results clearly show an advantage to one side or another.  Throughout our Frame Rating series of stories I have continued to tell the truth - that AMD cards are fantastic for single GPU configurations but need work on the multi-GPU side.  You can have your opinion obviously, but obviously we disagree.  As do many of the readers commenting here.

3. Sorry, I'm going to need more explanation on what you are saying here.  Frames are not produced in "batches" at all.  I think you are trying to describe the runt frame problem maybe?

September 18, 2013 | 09:59 PM - Posted by Allyn Malventano

2. Personally, I can't take your comment on a more serious manner. In your post, you state things that give me the impression that you don't fully understand reviews, or have poor views of NVidia graphics cards. At the very least, it is leading me to believe that you are bias towards AMD. Having favoritism, or a bias point of view to one company over the other, isn't a good way to approach a discussion or benchmark on any product.

Sucks how that works, doesn't it? Oh, for your point 3, it doesn't matter how fast a card can batch process *anything*, so long as what's presented to the user is inferior to the competition. The result is all that matters. Rolling back to point 1, your statements are moot as they are made without the far greater level of knowledge Ryan has - as he speaks with AMD about these various beta versions on an almost daily basis.

September 19, 2013 | 03:16 AM - Posted by technogiant (not verified)

As AMD have just stated that their Hawaii gpu's are smaller and more efficient but not intended to compete with the "ultra extreme" gpu's of Nvidia (aka 780/titan)as this is something that will be addressed by AMD's multi gpu cards....then it is all the more essential that AMD sorts these problems out properly and completely otherwise their product/business model is flawed as badly as their multi gpu performance.

September 19, 2013 | 03:24 AM - Posted by technogiant (not verified)

On a slightly different note but still regarding multi gpu I'd be interested in the views of you pcper guru's on the present state of multi gpu systems.

It's always seemed like such a waste to me using alternate frame rendering on multi gpu cards where each gpu has to have access to its own complete frame buffer size of memory.

Surely it would be better to use a tiled render approach where each gpu is working on individual tiles of the same frame and sharing one frame buffer sized chunk of memory?

September 26, 2013 | 05:20 AM - Posted by kn00tcn

someone always brings this up, & both ati & nv have said for years that AFR brings the most performance in the least complex way (unless a game engine has inter frame dependancies)

in the past, ati had tiled CF as an option, also scissor mode

but think about this, let's say you're doing 2 tiles at a horizontal split, you may end up with one card rendering an empty sky, the second rendering a ton of detail, basically resulting in a useless solution that doesnt scale

on top of that, you have to synchronize the tiles to display a final image at the same time, but the cards cant physically render at the exact same time, so you'll introduce lag or artifacts (which eyefinity does see)

i would say AFR is good enough & the way to go for multiple cards, but i would want to see a new paradigm... do you remember the first core2quad? it was 2 duals stitched together, imagine if 2 gpus were stitched together (no more mirroring the vram, just adjust the socket connections)

September 19, 2013 | 03:31 AM - Posted by Nvidia Shill (not verified)

September 19, 2013 | 09:13 AM - Posted by Ryan Shrout

LOL.  Some stories are funny, you know?

I replied to this here:

September 19, 2013 | 08:43 AM - Posted by Anonymous (not verified)

I don't even understand the point of this article.

Even before that article it was known Amd was going to fix it phases.

September 19, 2013 | 09:15 AM - Posted by Ryan Shrout

The point to showcase the very specific Eyefinity problems compared to Surround as they had not been discussed or shown in any form before today. 

September 19, 2013 | 11:18 AM - Posted by JJ White

Will you be taking a look at the Phanteks Enthoo Primo case? According to Hardware Canucks it might be the "Case of the year", not bad for such a small company entering the case market. I would be interested in what you think about it.

Here's the link to the HwC video:

September 19, 2013 | 03:47 PM - Posted by BIGGRIMTIM

How are your displays connected? I was having this issue until I connected all of my displays via DisplayPort. I know this is not ideal but it has eliminated the issue for me. I have 2 HD 7970s in crossfire and 3 Dell U2410 displays.

September 19, 2013 | 04:48 PM - Posted by Anonymous (not verified)

If the Asus PQ321 supports DisplayPort 1.2 and the HD 7970 supports DP 1.2 as well, and DP 1.2 can do 4k at 60Hz, then why is 4K necessarily a "dual head" affair? Is that simply due to the way the Asus was designed?

September 19, 2013 | 05:01 PM - Posted by Anonymous (not verified)

Ok. Nevermind. The whole tiled display thing. Is there a particular reason why 4k displays have to be tiled (or multi-headed)?

September 19, 2013 | 05:02 PM - Posted by Anonymous (not verified)

Ok. Nevermind. The whole tiled display thing. Is there a particular reason why 4k displays have to be tiled (or multi-headed)?

September 19, 2013 | 05:05 PM - Posted by Anonymous (not verified)

Ok. Nevermind. The whole tiled display thing.

From another comment on this site by NLPsajeeth:

"Currently there are no timing controllers that support 4K@60p. In order to drive the asus/sharp at 4K@60p, two separate TCONs are used. This is why this monitor has the unique capability of supporting dual HDMI. Each HDMI port feeds into its own TCON.

There is no 4K display that can do 60Hz without tiling. 4K@60p TCONs are supposed to start shipping in small amounts this year and in mass quantities in 2014."

September 19, 2013 | 07:22 PM - Posted by coffeefoot

Keep the faith, Ryan and co. Just continue to call it like you see it and let the chips fall where they may.
Hopefully AMD will get its stuff together otherwise they are going to lose a few folks.

September 19, 2013 | 09:42 PM - Posted by ArcRendition

I sincerely admire your journalistic integrity Ryan... as well everyone else at the PCper team!

-Stewart Graham

September 20, 2013 | 02:19 PM - Posted by ezjohny

what a difference in AMD graft, they improved on there driver.
Is this so with an APU + Graphic card. Good job Ryan.

September 20, 2013 | 09:42 PM - Posted by BigDaddyCF

Well currently rolling with 2x7970's on a 1920x1200 triple display setup. Can't say I ever really been personally bothered the various issues raised in the article in regards to the frame interlieaving and stepped tearing enough to stop playing, though I trust the guys over at PCPer to give it to me straight. I noticed the stuttering with crossfire more than anything else you guys brought up with your new testing methodology. I think most of us gamers at least gained a better understanding about the various issues involved. Sometimes my benchmarking applcation(be it FRAPS or Dxtory) would say I was getting a certain frame amount but the game just felt too jittery, whereas if I disabled crossfire the game felt more smooth even with a lower framerate.

That is not to say I haven't thoroughly enjoyed my 7970's/Eyefinity setup. When I've been been able to play at Eyefinity resolutions I've done so, when I haven't I've just adjusted my quality or resolution settings until I could get a smooth enough playing experience.

Do I hope that AMD is able to smooth out those circumstances where I can't play at a give resolution/quality due to micro-stuttering with crossfire, yeah that would be awesome. I think a lot of us out here still don't have a full appreciation for the phenomena due to not having been able to test multi-GPU solutions side by side, so it just comes down to "the game doesn't feel fluid enough at my current settings so I'll dial them down until it does", which I'm sure people have different sensitivities to. Keep up the good work PCPer crew.

September 21, 2013 | 09:11 PM - Posted by tbone8ty (not verified)

what about this ryan?

September 22, 2013 | 06:07 AM - Posted by Davros (not verified)

why use 2 hdmi cables when you can use a single displayport cable and the problem does not exist with displayport ?

September 22, 2013 | 07:04 AM - Posted by Russ (not verified)

At first I thought this article may have been over egging the problem with eyefinity + crossfire. Having now disconnected my second HD 7970 and played a few games in eyefinity I have seen that I is not. Radeon Pro may tell me that I'm getting half the FPS that I was but my eyes see the same low FPS experience.

Not impressed AMD, I feel like a chump for spending £300 on a card whose only additional effect to my system has been extra heat and noise.

Still at least I can go back and play Farcry 3 now with out the giant oversized HUD problem.

Thanks For the good article and thanks for bending AMD's ear.

September 22, 2013 | 07:14 AM - Posted by Gregster

A damned good read thanks Ryan. AMD owners should be pleased that these issues are highlighted and making sure AMD keep on their toes. Like the FCAT article, it was good to see AMD address the issue and get it fixed and again, it was PCper who made AMD aware of the issues (like they didn't already know!)and forced them into sorting that out for their users.

September 22, 2013 | 08:23 AM - Posted by drbaltazar (not verified)

I think a lot of hardware maker define CPU differently then can ask I wrote a bug report to and today 13.10 beta.if I recall message signal interrupt and its extended variant were implemented in vista for consumer?ROFL we know how vista was received so this might be one overlooked good case?in regedit MSI was enabled (sad was not for some reason ,can't enable it)but no amount of MSI set!(if it isn't set isn't it defaulting to one msi / socket?but I have 4 CPU in my i5 2500k(ya only physical CPU ms say)so imagine amd 8 core fx lol stuck with 1 MSI / msix.I think this is the cause.sadly on my system none were set . I normally tweak but from what I saw on ms it isn't a case of 0 or 1.and ms recommend hex value.Rolf a bit too complex for my knowledge.but you guys know a lot of hardcore tweaker . if I'm right ? I would be like what the eck am I the only one that used vista ?

September 22, 2013 | 08:33 AM - Posted by drbaltazar (not verified)

PS:what I wrote is for w8 64 bit!But I suspect a lot of hardware maker default to 1 (probably easier to implement)since socket come fro 2 to 12) detecting might be entertaining.

September 26, 2013 | 06:04 PM - Posted by drbaltazar (not verified) Resolution, Granularity, and Accuracy of System Time

Bottom line?I hate compromise!

Post new comment

The content of this field is kept private and will not be shown publicly.
  • Lines and paragraphs break automatically.
  • Allowed HTML tags: <a> <em> <strong> <cite> <code> <ul> <ol> <li> <dl> <dt> <dd> <blockquote><p><br>
  • Web page addresses and e-mail addresses turn into links automatically.

More information about formatting options

This question is for testing whether you are a human visitor and to prevent automated spam submissions.