NVIDIA Finally Launches GeForce GTX Titan Z Graphics Card

Subject: Graphics Cards | May 28, 2014 - 11:19 AM |
Tagged: titan z, nvidia, gtx, geforce

Though delayed by a month, today marks the official release of NVIDIA's Titan Z graphics card, the dual GK110 beast with the $3000 price tag. The massive card was shown for the first time in March at NVIDIA's GPU Technology Conference and our own Tim Verry was on the grounds to get the information

The details remain the same:

Specifically, the GTX TITAN Z is a triple slot graphics card that marries two full GK110 (big Kepler) GPUs for a total of 5,760 CUDA cores, 448 TMUs, and 96 ROPs with 12GB of GDDR5 memory on a 384-bit bus (6GB on a 384-bit bus per GPU). For the truly adventurous, it appears possible to SLI two GTX Titan Z cards using the single SLI connector. Display outputs include two DVI, one HDMI, and one DisplayPort connector.

The difference now of course is that all the clock speeds and pricing are official. 

titanzspecs.png

A base clock speed of 705 MHz with a Boost rate of 876 MHz places it well behind the individual GPU performance of a GeForce GTX 780 Ti or GTX Titan Black (rated at 889/980 MHz). The memory clock speed remains the same at 7.0 Gbps and you are still getting a massive 6GB of memory per GPU.

Maybe most interesting with the release of the GeForce GTX Titan Z is that NVIDIA seems to have completely fixated on non-DIY consumers with the card. We did not receive a sample of the Titan Z (nor did we get one of the Titan Black) and when I inquired as to why, NVIDIA PR stated that they were "only going to CUDA developers and system builders."

geforce-gtx-titan-z-3qtr.png

I think it is more than likely that after the release of AMD's Radeon R9 295X2 dual GPU graphics card on April 8th, with a price tag of $1500 (half of the Titan Z), the target audience was redirected. NVIDIA already had its eye on the professional markets that weren't willing to dive into the Quadro/Tesla lines (CUDA developers will likely drop $3k at the drop of a hat to get this kind of performance density). But a side benefit of creating the best flagship gaming graphics card on the planet was probably part of the story - and promptly taken away by AMD.

geforce-gtx-titan-z-bracket.png

I still believe the Titan Z will be an impressive graphics card to behold both in terms of look and style and in terms of performance. But it would take the BIGGEST NVIDIA fans to be able to pass up buying a pair of Radeon R9 295X2 cards for a single GeForce GTX Titan Z. At least that is our assumption until we can test one for ourselves.

I'm still working to get my hands on one of these for some testing as I think the ultra high end graphics card coverage we offer is incomplete without it. 

Several of NVIDIA's partners are going to be offering the Titan Z including EVGA, ASUS, MSI and Zotac. Maybe the most intersting though is EVGA's water cooled option!

evgatitanz.jpg

So, what do you think? Anyone lining up for a Titan Z when they show up for sale?

AMD Catalyst 14.6 Beta Driver Now Available, Adds Mixed Resolution Eyefinity

Subject: General Tech, Graphics Cards | May 27, 2014 - 12:00 AM |
Tagged: radeon, R9, R7, eyefinity, amd

AMD has just launched their Catalyst 14.6 Beta drivers for Windows and Linux. This driver will contain performance improvements for Watch Dogs, launching today in North America, and Murdered: Soul Suspect, which arrives next week. On Linux, the driver now supports Ubuntu 14.04 and its installation process has been upgraded for simplicity and user experience.

amd-146-eyefinity.png

Unless performance improvements are more important to you, the biggest feature is the support for Eyefinity with mixed resolutions. With Catalyst 14.6, you no longer need a grid of identical monitors. One example use case, suggested by AMD, is a gamer who purchases an ultra-wide 2560x1080 monitor. They will be able to add a pair of 1080p monitors on either side to create a 6400x1080 viewing surface.

amd-146-eyefinity2.png

If the monitors are very mismatched, the driver will allow users to letterbox to the largest rectangle contained by every monitor, or "expand" to draw the largest possible rectangle (which will lead to some assets drawing outside of any monitor). A third mode, fill, behaves like Eyefinity currently does. I must give AMD a lot of credit for leaving the choice to the user.

Returning to performance with actual figures, AMD claims "up to" 25% increases in Watch Dogs at 1080p or 28% at 1600p, compared to the previous version. The new CrossFire profile also claims up to 99% scaling in that game, at 2560x1600 with 8x MSAA. Murdered: Soul Suspect will see "up to" 16% improvements on a single card, and "up to" 93% scaling. Each of these results were provided by AMD, which tested on Radeon R9 290X cards. If these CrossFire profiles (well, first, are indicative of actual performance, and) see 99% scaling across two cards, that is pretty remarkable.

amd-146-jpeg.png

A brief mention, AMD has also expanded their JPEG decoder to Kabini. Previously, it was available to Kaveri, as of Catalyst 14.1. This allows using the GPU to display images, with their test showing a series of images being processed in about half of the time. While not claimed by AMD, I expect that the GPU will also be more power-efficient (as the processor can go back to its idle state much quicker, despite activitating another component to do so). Ironically, the three images I used for this news post are encoded in PNG. You might find that amusing.

AMD Catalyst 14.6 Beta Drivers should be now available at their download site.

Source: AMD

What if you can't afford that second R9 295X2?

Subject: Graphics Cards | May 26, 2014 - 05:16 PM |
Tagged: amd, radeon, r9 295x2, R9 290X

Through hard work or good luck you find yourself the proud owner of an R9 295X2 and a 4K display but somehow the performance just isn't quite good enough.  You can't afford another X2 though there is an R9 290X in your price range but you just aren't sure if it will help your system out at all.  That is where [H]ard|OCP steps in with this review where they prove that tri-fire in this configuration does indeed work.  Not only does it work, it allows you to vastly increase your performance over a 295X2 or to improve the performance somewhat while raising your graphics settings to new highs.  For those using 5760x1200 Eyefinity you probably already have your graphics options cranked; this upgrade will still offer you a linear increase in performance.  Not bad if you have the money to invest!

1399909966FHzxcUxVw3_1_3_l.jpg

"Will adding a single AMD Radeon R9 290X video card to the AMD Radeon R9 295X2 work? Will you get triple-GPU performance, ala TriFire CrossFire performance? This just might be a more financially feasible configuration for gamers versus QuadFire that provides a great gaming experience in Eyefinity and 4K resolutions."

Here are some more Graphics Card articles from around the web:

Graphics Cards

Source: [H]ard|OCP

NVIDIA Tegra K1 Benchmarks Spotted

Subject: General Tech, Graphics Cards, Mobile | May 22, 2014 - 04:58 PM |
Tagged: tegra k1, nvidia, iris pro, iris, Intel, hd 4000

The Chinese tech site, Evolife, acquired a few benchmarks for the Tegra K1. We do not know exactly where they got the system from, but we know that it has 4GB of RAM and 12 GB of storage. Of course, this is the version with four ARM Cortex-A15 cores (not the upcoming, 64-bit version based on Project Denver). On 3DMark Ice Storm Unlimited, it was capable of 25737 points, full system.

nvidia-k1-benchmark.jpg

Image Credit: Evolife.cn

You might remember that our tests with an Intel Core i5-3317U (Ivy Bridge), back in September, achieved a score of 25630 on 3DMark Ice Storm. Of course, that was using the built-in Intel HD 4000 graphics, not a discrete solution, but it still kept up for gaming. This makes sense, though. Intel HD 4000 (GT2) graphics has a theoretical performance of 332.8 GFLOPs, while the Tegra K1 is rated at 364.8 GFLOPs. Earlier, we said that its theoretical performance is roughly on par with the GeForce 9600 GT, although the Tegra K1 supports newer APIs.

Of course, Intel has released better solutions with Haswell. Benchmarks show that Iris Pro is able to play Battlefield 4 on High settings, at 720p, with about 30FPS. The HD 4000 only gets about 12 FPS with the same configuration (and ~30 FPS on Low). This is not to compare Intel to NVIDIA's mobile part, but rather compare Tegra K1 to modern, mainstream laptops and desktops. It is getting fairly close, especially with the first wave of K1 tablets entering at the mid-$200 USD MSRP in China.

As a final note...

There was a time where Tim Sweeney, CEO of Epic Games, said that the difference between high-end and low-end PCs "is something like 100x". Scaling a single game between the two performance tiers would be next-to impossible. He noted that ten years earlier, that factor was more "10x".

Now, an original GeForce Titan is about 12x faster than the Tegra K1 and they support the same feature set. In other words, it is easier to develop a game for the PC and high-end tablet than it was to develop an PC game for high-end and low-end machines, back in 2008. PC Gaming is, once again, getting healthier.

Source: Evolife.cn

AMD Radeon R9 Prices Have Leveled Out, R9 280 Drops to $249

Subject: Graphics Cards | May 15, 2014 - 06:16 PM |
Tagged: radeon, R9 290X, r9 290, r9 280x, r9 280, amd

Just the other day AMD sent out an email to the media to discuss the current pricing situation of the Radeon R9 series of graphics cards. This email started with the following statement.

You’ve seen many articles, discussions online about the AMD Radeon™ R9 lineup – especially chatter about pricing and availability. As we’ve talked about it before, the demand for the R9 lineup has been nothing but astonishing, and went well beyond our most optimistic expectations. That created a situation where gamers weren’t able to purchase their desired R9 graphics card.

Radeon_R9Fan_WaterCooler_Product_Shot_Birdseye_5in300dpi_4c.jpg

Clearly AMD would not bring up the subject if the current situation was BAD news so guess what? All seems to be back normal (or expected) in terms of AMD Radeon R9 pricing and card availability. Take a look at the table below to get an idea of where Radeon's currently stand.

  Amazon.com Newegg.com
Radeon R9 295X2 $1524 $1499
Radeon R9 290X  $549 $529
Radeon R9 290 $379 $399
Radeon R9 280X $289 $299
Radeon R9 280 $249 $249
Radeon R9 270X $199 $189
Radeon R9 270 $169 $179

There is one price change that differs from the products' launch - the SEP of the Radeon R9 280 has dropped from $279 to $249. Nothing dramatic but a nice change.

Maybe most interesting is this line from the AMD email.

Now that product is available and at suggested pricing, these prices will remain stable. No more madness like you saw in Q1.

That emphasis is AMD's. I'm not quite sure how the company thinks they can keep a tight control on pricing now if it wasn't able to do so before, but more than likely, with the rush for coin mining hardware somewhat dying off, the prediction will hold true. (As a side note, there appears to be some discounts to be found on used Radeon hardware these days...)

AMD-Never-Settle-Forever-2014-01.jpg

Of course the AMD bundling promotion known as Never Settle Forever is still going strong with these new prices as well. Scott wrote up a story detailing this latest incarnation of the promotion and he and I both agree that while free is always good great, the age of most of the titles in the program is a bit of a problem. But AMD did note in this email that they have "lined up a few brand new games to add to this promotion, and they'll [sic] be sharing more info with you in the next few weeks!"

Xiaomi MiPad Tablet is Tegra K1 Powered

Subject: General Tech, Graphics Cards, Processors, Mobile | May 15, 2014 - 05:02 PM |
Tagged: nvidia, xaiomi, mipad, tegra k1

Tegra K1 is NVIDIA's new mobile processor and this first to implement the Kepler graphics architecture. In other words, it has all of the same graphics functionality as a desktop GPU with 364 GigaFLOPs of performance (a little faster than a GeForce 9600 GT). This is quite fast for a mobile product. For instance, that amount of graphics performance could max out Unreal Tournament 3 to 2560x1600 and run Crysis at 720p. Being Kepler, it supports OpenGL 4.4, OpenGL ES 3.1, DirectX 11 and 12, and GPU compute languages.

Xiaomi is launching their MiPad in Beijing, today, with an 8-inch 2048x1536 screen and the Tegra K1. They will be available in June (for China) starting at $240 USD for the 16GB version and going up to $270 for the 64GB version. Each version has 2GB of RAM, an 8MP rear-facing camera, and a 5MP front camera.

Now, we wait and see if any Tegra K1 devices come to North America and Europe - especially at that price point.

Source: NVIDIA

NVIDIA Titan Z Missed Its Release Date

Subject: General Tech, Graphics Cards | May 12, 2014 - 08:00 PM |
Tagged: titan z, nvidia, gtx titan z, geforce

To a crowd of press and developers at their GTC summit, NVIDIA announced the GeForce GTX Titan Z add-in board (AIB). Each of the two, fully unlocked, GK110 GPUs would each have access to 6GB of GDDR5 memory (12GB total). The card was expected to be available on May 8th but has yet to surface. As NVIDIA has yet to comment on the situation, many question whether it ever will.

nvidia-titan-z-where.png

And then we get what we think are leaked benchmarks (note: two pictures).

One concern about the Titan Z was its rated 8 TeraFLOPs of compute performance. This is a fairly sizable reduction from the theoretical maximum of 10.24 TeraFLOPs of two Titan Black processors and even less than two first-generation Titans (9 TeraFLOPs combined). We expected that this is due to reduced clock rates. What we did not expect is for benchmarks to show the GPUs boost way above those advertised levels, and even beyond the advertised boost clocks of the Titan Black and the 780 Ti. The card was seen pushing 1058 MHz in some sections, which leads to a theoretical compute performance of 12.2 TeraFLOPs (6.1 TeraFLOPs per GPU) in single precision. That is a lot.

These benchmarks also show that NVIDIA has a slight lead over AMD's R9 295X2 in many games, except Battlefield 4 and Sleeping Dogs (plus 3DMark and Unigine). Of course, these benchmarks measure the software reported frame rate and frame times and those may or may not be indicative of actual performance. While I would say that the Titan Z appears to have a slight performance lead over the R9 295X2, although a solid argument for an AMD performance win exists, it does so double the cost (at its expected $3000 USD price point). That is not up for debate.

Whichever card is faster, AMD's is half the price and available for purchase right now.

So, until NVIDIA says anything, the Titan Z is in limbo. I am sure there exists CUDA developers who await its arrival. Personally, I would just get three Titan Blacks since you are going to need to manually schedule your workloads across multiple processors anyway (or 780 Tis if 32-bit arithmetic is enough precision). That is, of course, unless you cannot physically fit enough GeForce Titan Blacks in your motherboard and, as such, you require two GK110 chips per AIB (but not enough to bother writing a cluster scheduling application).

Source: Unknown

NVIDIA GTX 780 Ti on Thunderbolt 2 by DIYers

Subject: General Tech, Graphics Cards, Mobile | May 7, 2014 - 02:26 AM |
Tagged: Thunderbolt 2, thunderbolt, nvidia, GeForce GTX 780 Ti

Externally-attached GPUs have been a topic for many years now. Numerous companies have tried, including AMD and Lucid, but no solution has ever been a widely known and available product. Even as interfaces increase in bandwidth and compatibility with internal buses, it has never been something that a laptop salesperson could suggest to users who want to dock into a high-performance station at home. At best, we are seeing it in weird "coin mining" racks to hang way more GPUs above a system than could physically mount on the motherboard.

Apparently that has not stopped the DIY community, according to chatter on Tech Inferno forums. While the above video does not really show the monitor, MacBook Pro, and GPU enclosure at the same time, let alone all wired together and on, it seems reasonable enough. The video claims to give the MacBook Pro (running Windows 8.1) access to a GeForce GTX 780 Ti with fairly high performance, despite the reduced bandwidth. Quite cool.

Check out the forum post for a few more details.

Source: Tech Inferno

PowerColor Teasing Dual GPU Graphics Card With Massive Air Cooler

Subject: Graphics Cards | May 6, 2014 - 03:36 AM |
Tagged: r9 295x2, powercolor, hawaii, dual gpu, devil 13

PowerColor has been teasing a new graphics card on its Facebook page. The photos show a macro shot of the Devil 13 logo along with captions hitting at the new card being a dual GPU monster including one caption referring the upcoming Devil 13 as a "dual beast."

PowerColor's previous Devil 13 branded graphics card was the Radeon HD 7990 Devil 13 which contained two HD 7970 "Tahiti" GPUs on one PCB. Coincidentally, AMD recently launched a new dual GPU reference design based around two R9 290x "Hawaii" GPUs called the R9 295x2. It is still rumor and speculation at this point, but the timing and leaked photos seem to point squarely at the upcoming Devil 13 card being the first air cooled custom R9 295x2!

PowerColor Dual Beast R9 295X2 Dual Hawaii GPU Air Cooled.jpg

Adding credence to the rumors, leaked photos have appeared online with a PCB backplate that appears to match the backplate shown in the official teaser photo. The leaked photos show an absolutely beastly triple slot graphics card that places two GPUs in CrossFire on a single custom PCB powered by four 8-pin PCI-E power connectors and cooled by a gargantuan HSF comprised of an aluminum fin stack and multiple large diameter copper heatpipes along with three fans. The cooler and PCB are reinforced with brackets and a metal backplate to help keep the air cooler in pace and the PCB from bending.

PowerColor Devil 13 Dual GPU Air Cooled Graphics Card 4 8-pin PCIe.jpg

If the rumors hold true, PowerColor will be unveiling the first air cooled dual GPU R9 295X2 graphics card which is an impressive feat of engineering! Using four 8-pin PCI-E power connectors definitely suggests that aftermarket overclocking is encouraged and supported even if PowerColor does not end up factory overclocking their dual GPU beast.

For reference, the stock AMD R9 295X2 features two full Hawaii GPUs with 5,632 stream processors clocked at up to 1018 MHz interfaced with 8GB of total GDDR5 memory over a 512-bit bus (each GPU has 4GB of memory and a 512-bit bus). AMD rates this configuration at 11.5 TFLOPS of single precision performance. The reference R9 295X2 has a 500W TDP and uses two 8-pin PCI-E power connectors.

Please excuse me while I wipe the drool off of my keyboard...

Stay tuned to PC Perspective for more details on the mysterious dual GPU Devil 13 from PowerColor!

In the meantime, check out our full review of the R9 295X2 (and the Hawaii architecture) and what happens when you put two R9 295X2s in Quad CrossFire into a single system for 4K gaming goodness!

Source: PowerColor

GeForce Experience 2.0.1 Update Released

Subject: General Tech, Graphics Cards | May 5, 2014 - 05:03 PM |
Tagged: nvidia, geforce experience, shield

NVIDIA has released version 2.0.1 of GeForce Experience. This update does not bring many new features, hence why it is a third-level increment to the version number, but is probably worthwhile to download regardless. Its headlining feature is security enhancements with OpenSSL under remote GameStream on SHIELD. The update also claims to improve streaming quality and reduce audio latency.

nvidia-shield-gamestream-02.jpg

While they do not seem to elaborate, I assume this is meant to fix Heartbleed, which is an exploit that allows an attacker to receive a small snapshot of active memory. If that is that case, it is unclear whether the SHIELD, the host PC during a game session, or both endpoints are affected.

The new GeForce Experience is available at the NVIDIA website. If it is running, it will also ask you to update it, of course.

Source: NVIDIA

Asus Launches GTX TITAN Z Dual GK110 Graphics Card

Subject: Graphics Cards | May 2, 2014 - 01:29 AM |
Tagged: titan z, nvidia, gpgpu, gk110, dual gpu, asus

NVIDIA unveiled the GeForce GTX TITAN Z at the GPU Technology Conference last month, and the cards will be for sale soon from various partners. ASUS will be one of the first AIB partners to offer a reference TITAN-Z.

The ASUS GTX TITAN Z pairs two full GK110-based GPUs with 12GB of GDDR5 memory. The graphics card houses a total of 5,760 CUDA cores, 480 texture manipulation units (TMUs), and 96 ROPs. Each GK110 GPU interfaces with 6GB of GDDR5 memory via a 384-bit bus. ASUS is using reference clockspeeds with this card, which means 705 MHz base and up to 876 MHz GPU Boost for the GPUs and 7.0 GHz for the memory.

ASUS GTX TITAN Z Dual GPU Graphics Card.jpg

For comparison, the dual-GPU TITAN Z is effectively two GTX TITAN Black cards on a single PCB. However, the TITAN Black runs at 889 MHz base and up to 980 MHz GPU Boost. A hybrid water cooling solution may have allowed NVIDIA to maintain the clockspeed advantage, but doing so would compromise the only advantage the TITAN Z has over using two (much cheaper) TITAN Blacks in a workstation or server: card density. A small hit in clockspeed will be a manageable sacrifice for the target market, I believe.

The ASUS GTX TITAN Z has a 375W TDP and is powered by two 8-pin PCI-E power connectors. The new flagship dual GPU NVIDIA card has an MSRP of $3,000 and should be available in early May.

Source: Asus

AMD Mantle Private Beta Announced

Subject: General Tech, Graphics Cards | May 1, 2014 - 08:00 AM |
Tagged: Mantle, amd

As our readers are well aware, Mantle is available for use with a few games. Its compatibility begun with the beta Catalyst 14.1 driver and an update for Battlefield 4. AMD was quite upfront about the technology, even granting a brief interview with Guennadi Riguer, Chief Architect of the API to fill in a few of the gaps left from their various keynote speeches.

AMD_Mantle_Logo.png

What is under lock and key, however, is the actual software development kit (SDK). AMD claimed that it was too immature for the public. It was developed in partnership with DICE, Oxide Games, and other, established developers to fine-tune its shape, all the while making it more robust. That's fine. They have a development plan. There is nothing wrong with that. Today, while the SDK is still not public and sealed by non-disclosure agreement, AMD is accepting applications from developers who are requesting to enter the program.

If you want to develop a Mantle application or game, follow the instructions at their website for AMD to consider you. They consider it stable, performant, and functional enough for "a broader audience in the developer community".

AMD cites 40 developers already registered, up from seven (DICE, Crytek, Oxide, etc.).

If you are not a developer, then this news really did not mean too much to you -- except that progress is being made.

Source: AMD

Post Tax Day Celebration! Win an EVGA Hadron Air and GeForce GTX 750!

Subject: Editorial, General Tech, Graphics Cards | April 30, 2014 - 10:05 AM |
Tagged: hadron air, hadron, gtx 750, giveaway, evga, contest

Congrats to our winner: Pierce H.! Check back soon for more contests and giveaways at PC Perspective!!

In these good old United States of America, April 15th is a trying day. Circled on most of our calendars is the final deadline for paying up your bounty to Uncle Sam so we can continue to have things like freeway systems and universal Internet access. 

But EVGA is here for us! Courtesy of our long time sponsor you can win a post-Tax Day prize pack that includes both an EVGA Hadron Air mini-ITX chassis (reviewed by us here) as well as an EVGA GeForce GTX 750 graphics card. 

evgacontestapril.jpg

Nothing makes paying taxes better than free stuff that falls under the gift limit...

With these components under your belt you are well down the road to PC gaming bliss, upgrading your existing PC or starting a new one in a form factor you might not have otherwise imagined. 

Competing for these prizes is simple and open to anyone in the world, even if you don't suffer the same April 15th fear that we do. (I'm sure you have your own worries...)

  1. Fill out the form at the bottom of this post to give us your name and email address, in addition to the reasons you love April 15th! (Seriously, we need some good ideas for next year to keep our heads up!) Also, this does not mean you should leave a standard comment on the post to enter, though you are welcome to do that too.
     
  2. Stop by our Facebook page and give us a LIKE (I hate saying that), head over to our Twitter page and follow @pcper and heck, why not check our our many videos and subscribe to our YouTube channel?
     
  3. Why not do the same for EVGA's Facebook and Twitter accounts?
     
  4. Wait patiently for April 30th when we will draw and update this news post with the winners name and tax documentation! (Okay, probably not that last part.)

A huge thanks goes out to friends and supporters at EVGA for providing us with the hardware to hand out to you all. If it weren't for sponsors like this PC Perspective just couldn't happen, so be sure to give them some thanks when you see them around the In-tar-webs!!

Good luck!

Source: EVGA

NVIDIA Announces Watch_Dogs Bundle with GeForce GPUs

Subject: Graphics Cards | April 29, 2014 - 10:22 AM |
Tagged: nvidia, watch_dogs, watch dogs, bundle, geforce

A bit of a surprise email found its way to my inbox today that announced NVIDIA's partnership with Ubisoft to include copies of the upcoming Watch_Dogs game with GeForce GTX graphics cards. 

watchdogs.jpg

Gamers that purchase a GeForce GTX 780 Ti, GTX 780, GTX 770 or GTX 760 from select retailers will qualify for a free copy of the game. You can details on this bundle and available GPUs to take advantage of it at Amazon.com!

The press release also confirms inclusion of NVIDIA exclusive features like TXAA and HBAO+ in the game itself, which is interesting. From what I am hearing, Watch_Dogs is going to be a beast of a game on GPU hardware and we are looking forward to using it as a test platform going forward.

Full press release is included below.

OWN THE TECH AND CONTROL THE CITY WITH NVIDIA® AND UBISOFT®

Select GeForce GTX GPUs Now Include the Hottest Game of the Year: Watch_Dogs™

Santa Clara, CA  April 29, 2014 — Destructoid calls it one of the “most wanted games of 2014.” CNET said it was “one of the most anticipated games in recent memory.” MTV said it’s one of the “Can’t-Miss Video Games of 2014.” This, all before anyone out there has even played it.

So, starting today(1), gamers who purchase select NVIDIA® GeForce® GTX® 780 Ti, 780, 770 and 760 desktop GPUs can get their chance to play Watch_Dogs™, the new PC game taking the world by storm and latest masterpiece from Ubisoft®.

Described as a “polished, refined and truly next generation experience,” in Watch_Dogs you play as Aiden Pearce, a brilliant hacker whose criminal past led to a violent family tragedy. While seeking justice, you will monitor and hack those around you, access omnipresent security cameras, download personal information to locate a target, control traffic lights and public transportation to stop the enemy and more.

Featuring NVIDIA TXAA and HBAO+ technology for an interactive, immersive experience, it’s clear that gamers can’t wait to play Watch_Dogs, especially considering the effusive praise that the official trailer received. Launched mere weeks ago, the trailer has already been viewed more than a combined 650,000 times. For gamers, Watch_Dogs seamlessly blends a mixture of single-player and multiplayer action in a way never before seen, and Ubisoft has gone one step further in creating a unique ctOS mobile companion app for users of smartphone and tablet devices allowing for even greater access to the fun. If you haven’t checked out the trailer, please check it out here: https://www.youtube.com/watch?v=3eHCJ8pWdf0.

The GeForce GTX and Watch_Dogs bundle is available starting today from leading e-tailers including Newegg, Amazon.com, TigerDirect, NCIX; add-in card vendors such as EVGA; and nationwide system builders including AVADirect, CyberPower, Digital Storm, Falcon Northwest, iBUYPOWER, Maingear, Origin PC, Puget Systems, V3 Gaming PC and Velocity Micro. For a full list of participating partners, please visit: www.GeForce.com/GetWatchDogs.

Source: NVIDIA

Another GPU Driver Showdown: AMD vs NVIDIA in Linux

Subject: General Tech, Graphics Cards | April 27, 2014 - 04:22 AM |
Tagged: nvidia, linux, amd

GPU drivers have been a hot and sensitive topic at the site, especially recently, probably spurred on by the announcements of Mantle and DirectX 12. These two announcements admit and illuminate (like a Christmas tree) the limitations of APIs on gaming performance. Both AMD and NVIDIA have their recent successes and failures on their respective fronts. This will not deal with that, though. This is a straight round-up of new GPUs running the latest drivers... in Linux.

7-TuxGpu.png

Again, results are mixed and a bit up for interpretation.

In all, NVIDIA tends to have better performance with its 700-series parts than equivalently-priced R7 or R9 products from AMD, especially in low-performance Source Engine titles such as Team Fortress 2. Sure, even the R7 260X was almost at 120 FPS, but the R9 290 was neck-and-neck with the GeForce GTX 760. The GeForce GTX 770, about $50 cheaper than the R9 290, had a healthy 10% lead over it.

In Unigine Heaven, however, the AMD R9 290 passed the NVIDIA GTX 770 by a small margin, coming right in line with it's aforementioned $50-bigger price tag. In that situation, where performance became non-trivial, AMD caught up (but did not beat). Also, third-party driver support is more embraced by AMD than NVIDIA. On the other hand, NVIDIA's proprietary drivers are demonstrably better, even if you would argue that the specific cases are trivial because of overkill.

And then there's Unvanquished, where AMD's R9 290 did not achieve triple-digit FPS scores despite the $250 GTX 760 getting 110 FPS.

Update: As pointed out in the comments, some games perform significantly better on the $130 R7 260X than the $175 GTX 750 Ti (HL2: Lost Coast, TF2, OpenArena, Unigine Sanctuary). Some other games are the opposite, with the 750 Ti holding a sizable lead over the R7 260X (Unigine Heaven and Unvanquished). Again, Linux performance is a grab bag between vendors.

There's a lot of things to consider, especially if you are getting into Linux gaming. I expect that it will be a hot topic, soon, as it picks up... ... Steam.

Source: Phoronix

AMD Catalyst 14.4 Release Candidate is now available

Subject: Graphics Cards | April 22, 2014 - 01:06 PM |
Tagged: catalyst 14.4, catalyst, amd

The latest available AMD Catalyst Windows and Linux drivers can be found here:
AMD Catalyst Windows: http://support.amd.com/en-us/kb-articles/Pages/latest-catalyst-windows-beta.aspx
AMD Catalyst Linux: http://support.amd.com/en-us/kb-articles/Pages/latest-linux-beta-driver.aspx

image001.jpg

Highlights of AMD Catalyst™ 14.4 Windows Driver

  • Support for the AMD Radeon R9 295X

CrossFire fixes enhancements:

  • Crysis 3 – frame pacing improvements
  • Far Cry 3 – 3 and 4 GPU performance improvements at high quality settings, high resolution settings
  • Anno 2070 – Improved CrossFire scaling up to 34%
  • Titanfall – Resolved in game flickering with CrossFire enabled
  • Metro Last Light – Improved Crossfire scaling up to 10%
  • Eyefinity 3x1 (with three 4K panels) no longer cuts off portions of the application
  • Stuttering has been improved in certain applications when selecting mid-Eyefinity resolutions with V-sync Enabled

Full support for OpenGL 4.4
Mantle beta driver improvements:

  • BattleField 4: Performance slowdown is no longer seen when performing a task switch/Alt-tab
  • BattleField 4: Fuzzy images when playing in rotated SLS resolution with an A10 Kaveri system

Highlights of AMD Catalyst™ 14.1 Linux Driver

  • Support for the AMD Radeon R9 295X
  • Ubuntu 12.04.4 support
  • Full support for OpenGL 4.4

Resolved Issue highlights:

  • Corruption and system hang observed while running Sanctuary BM with Tear Free Desktop enabled
  • Memory leak about hardware context EGL create context error for glesx
  • GPU hand in CrossFire Mode [Piglit]
  • Test "spec/arb_vertex_array_object" failed [Piglit]
  • Test "glx/GLX_EXT_import_context/free context" failed [Piglit]
  • Test "spec/ARB_seamless_cube_map" failed Piglit]
  • Test "texture swizzle with border color" failed
  • Glxtest failures observed in log file Blank screen observed while running steam games with Big picture
  • 4ms delay observed in the glxSwapBuffers when vsync is enabled
  • RBDoom3BFG the game auto quit when use the security camera terminal
  • ETQW segmentation fault

Source: AMD

Nope, Never Settling... Forever. More Bundles.

Subject: General Tech, Graphics Cards | April 21, 2014 - 01:55 PM |
Tagged: radeon, never settle forever, never settle, amd

AMD has been taking PC gaming very seriously, especially over the last couple of years. While they have a dominant presence in the console space, with only IBM in opposition, I believe that direct licensing revenue was not their main goal, rather that they hope to see benefits carry over to the PC and maybe mobile spaces, eventually. In the PC space, Never Settle launched as a very successful marketing campaign. While it had a stutter with the launch of the R9 (and R7) product lines, it is back and is still called, "Never Settle Forever".

AMD-Never-Settle-Forever-2014-01.jpg

Keeping with Forever's alteration to the Never Settle formula, the type of card that you purchase yields a Gold, Silver, or Bronze reward. Gold (the R9 280 and R9 290 series, and the R9 295X2) gets three free games in the Gold tier, Silver (R9 270 and R7 260 series) gets two in the Silver tier, and Bronze (R7 250 and R7 240 series) gets one free game in the Bronze tier. By and large, the tiers are the same as last time plus a few old games and one upcoming Square Enix release: Murdered: Soul Suspect. They have also made deals with certain independent developers, where two indie titles bundled together count as one choice.

The complete breakdown of games is as follows:

 
Gold
(Choose 3)
Silver
(Choose 2)
Bronze
(Choose 1)
Murdered: Soul Suspect (June 3, 2014) Yes Yes No
Thief Yes Yes No
Tomb Raider Yes Yes No
Hitman: Absolution Yes Yes No
Sleeping Dogs Yes Yes No
Dungeon Siege III Yes Yes Yes
Dirt 3 Yes Yes Yes
Alan Wake Yes Yes Yes
Darksiders Yes Yes Yes
Darksiders II Yes Yes Yes
Company of Heroes 2 Yes Yes Yes
Total War: Shogun 2 Yes Yes Yes
Titan Quest (Gold Edition) Yes Yes Yes
Supreme Commander (Gold Edition) Yes Yes Yes
Deus Ex: Human Revolution Yes Yes No
Payday 2 Yes Yes No
Just Cause 2 Yes Yes Yes
Banner Saga + Mutant Blobs Attack (indie combo) Yes Yes Yes
Guacamelee + DYAD (indie combo) Yes Yes Yes
Mutant Blobs Attack + DYAD (indie combo) Yes Yes Yes
Banner Saga + DYAD (indie combo) Yes Yes Yes
Mutant Blobs Attack + Guacamelee (indie combo) Yes Yes Yes

Oddly enough, there does not seem to be a Banner Saga + Guacamelee combo...

... the only impossible combination.

AMD has also announced that Never Settle will continue for more "additions" in 2014. Which ones? Who knows. It is clear that they have a great working relationship with Square Enix Europe, including basically their last six major titles in Never Settle and keeping them there, but there is not really anything from them on the horizon (at least, not announced). AMD does sound confident in having other deals lined up this year, however.

amd-never-settle-forever-2014-02.jpg

Never Settle Forever graphics cards are available now "at participating retailers". Bundle codes can be redeemed any time between now and August 31st.

There is some regional variance in game availability, however. Read up before you purchase (especially if you live in Japan). You should be fine if you live in North America, Europe, Middle East, Africa, New Zealand, Australia, and Latin America, though, at least where AMD products are available. Still, it is a good idea to check.

Source: AMD

An overclocked flagship GPU duel

Subject: Graphics Cards | April 17, 2014 - 04:10 PM |
Tagged: amd, nividia, gigabyte, asus, R9 290X, GeForce GTX 780 Ti, factory overclocked

In the green trunks is the ASUS GTX 780 Ti DirectCU II OC which [H]ard|OCP overclocked to the point they saw in game performance of 1211MHz GPU and 7.2GHz on the memory.  In the red trunks we find Gigabyte's R9 290X 4GB OC weighing in at 1115MHz and 5.08GHz for the GPU and memory respectively.  Both cards have been pushed beyond the factory overclock that they came with and will fight head to head in such events as Battling the Field, Raiding the Tomb and counting to three twice, once in a Crysis and again in a Far Cry from safety.  Who will triumph?  Will the battle be one sided or will the contenders trade top spot depending on the challenge?  Get the full coverage at [H]ard|OCP!

1397411267OPl8cM2MpM_2_8_l.jpg

"Today we look at the GIGABYTE R9 290X 4GB OC and ASUS GeForce GTX 780 Ti DirectCU II OC video cards. Each of these video cards features a custom cooling system, and a factory overclock. We will push the overclock farther and put these two video cards head-to-head for a high-end performance comparison."

Here are some more Graphics Card articles from around the web:

Graphics Cards

Source: [H]ard|OCP

Win a Galaxy GeForce GTX 750 Ti GC or GeForce GTX 750 GC!

Subject: General Tech, Graphics Cards | April 16, 2014 - 11:39 AM |
Tagged: video, giveaway, galaxy, contest

UPDATE: Our winners have been selected and notified! Thanks to everyone for participating and stayed tuned to pcper.com as we'll have more contests starting VERY SOON!!!

Our sponsors are the best, they really are. Case in point - Galaxy would like us to give away a pair of graphics cards to our fans. On the block for the contest are a Galaxy GTX 750 Ti GC and a Galaxy GTX 750 GC option, both based on the latest generation Maxwell GPU architecture from NVIDIA.

I posted a GTX 750 Ti Roundup story that looked at the Galaxy GTX 750 Ti GC option and it impressed in both stock performance and in the amount of overclocking headroom provided by the custom cooler.

IMG_9862.JPG

How can you win these awesome prizes? Head over to our YouTube channel to find or just watch the video below! You need to be a subscriber to our YouTube channel as well as leave a comment on the video itself over on YouTube.

Anyone, any where in the world can win. We'll pick a winner on April 16th - good luck!

Source: YouTube

NVIDIA GeForce Driver 337.50 Early Results are Impressive

Subject: Graphics Cards | April 11, 2014 - 03:30 PM |
Tagged: nvidia, geforce, dx11, driver, 337.50

UPDATE: We have put together a much more comprehensive story based on the NVIDIA 337.50 driver that includes more cards and more games while also disputing the Total War: Rome II results seen here. Be sure to read it!!

When I spoke with NVIDIA after the announcement of DirectX 12 at GDC this past March, a lot of the discussion centered around a pending driver release that promised impressive performance advances with current DX11 hardware and DX11 games. 

What NVIDIA did want to focus on with us was the significant improvements that have been made on the efficiency and performance of DirectX 11.  When NVIDIA is questioned as to why they didn’t create their Mantle-like API if Microsoft was dragging its feet, they point to the vast improvements possible and made with existing APIs like DX11 and OpenGL. The idea is that rather than spend resources on creating a completely new API that needs to be integrated in a totally unique engine port (see Frostbite, CryEngine, etc.) NVIDIA has instead improved the performance, scaling, and predictability of DirectX 11.

09.jpg

NVIDIA claims that these fixes are not game specific and will improve performance and efficiency for a lot of GeForce users. Even if that is the case, we will only really see these improvements surface in titles that have addressable CPU limits or very low end hardware, similar to how Mantle works today.

Lofty goals to be sure. This driver was released last week and I immediately wanted to test and verify many of these claims. However, a certain other graphics project kept me occupied most of the week and then a short jaunt to Dallas kept me from the task until yesterday. 

To be clear, I am planning to look at several more games and card configurations next week, but I thought it was worth sharing our first set of results. The test bed in use is the same as our standard GPU reviews.

Test System Setup
CPU Intel Core i7-3960X Sandy Bridge-E
Motherboard ASUS P9X79 Deluxe
Memory Corsair Dominator DDR3-1600 16GB
Hard Drive OCZ Agility 4 256GB SSD
Sound Card On-board
Graphics Card NVIDIA GeForce GTX 780 Ti 3GB
NVIDIA GeForce GTX 770 2GB
Graphics Drivers NVIDIA: 335.23 WHQL, 337.50 Beta
Power Supply Corsair AX1200i
Operating System Windows 8 Pro x64

The most interesting claims from NVIDIA were spikes as high as 70%+ in Total War: Rome II, so I decided to start there. 

First up, let's take a look at the GTX 780 Ti SLI results, the flagship gaming card from NVIDIA.

TWRome2_2560x1440_OFPS.png

TWRome2_2560x1440_PER.png

TWRome2_2560x1440_PLOT.png

With this title, running at the Extreme preset, jumps from an average frame rate of 59 FPS to 88 FPS, an increase of 48%! Frame rate variance does increase a bit with the faster average frame rate but it stays within limits of smoothness, but barely.

Next up, the GeForce GTX 770 SLI results.

TWRome2_2560x1440_OFPS.png

TWRome2_2560x1440_PER.png

TWRome2_2560x1440_PLOT.png

Results here are even more impressive as the pair of GeForce GTX 770 cards running in SLI jump from 29.5 average FPS to 51 FPS, an increase of 72%!! Even better, this occurs without any kind of frame rate variance increase and in fact, the blue line of the 337.50 driver is actually performing better in that perspective.

All of these tests were run with the latest patch on Total War: Rome II and I did specifically ask NVIDIA if there were any differences in the SLI profiles between these two drivers for this game. I was told absolutely not - this just happens to be the poster child example of changes NVIDIA has made with this DX11 efficiency push.

Of course, not all games are going to see performance improvements like this, or even improvements that are measurable at all. Just as we have seen with other driver enhancements over the years, different hardware configurations, image quality settings and even scenes used to test each game will shift the deltas considerably. I can tell you already that based on some results I have (but am holding for my story next week) performance improvements in other games are ranging from <5% up to 35%+. While those aren't reaching the 72% level we saw in Total War: Rome II above, these kinds of experience changes with driver updates are impressive to see.

Even though we are likely looking at the "best case" for NVIDIA's 337.50 driver changes with the Rome II results here, clearly there is merit behind what the company is pushing. We'll have more results next week!