Subject: General Tech, Graphics Cards | May 28, 2014 - 01:49 PM | Jeremy Hellstrom
Tagged: benchmarking, 3dmark
HELSINKI, FINLAND – May 28, 2014 – Futuremark today announced 3DMark Sky Diver, a new DirectX 11 benchmark test for gaming laptops and mid-range PCs. 3DMark Sky Diver is the ideal test for benchmarking systems with mainstream DirectX 11 graphics cards, mobile GPUs, or integrated graphics. A preview trailer for the new benchmark shows a wingsuited woman skydiving into a mysterious, uncharted location. The scene is brought to life with tessellation, particles and advanced post-processing effects. Sky Diver will be shown in full at Computex from June 3-7, or find out more on the Futuremark website.
Jukka Mäkinen, Futuremark CEO said, "Some people think that 3DMark is only for high-end hardware and extreme overclocking. Yet millions of PC gamers rely on 3DMark to choose systems that best balance performance, efficiency and affordability. 3DMark Sky Diver complements our other tests by providing the ideal benchmark for gaming laptops and mainstream PCs."
3DMark - The Gamer's Benchmark for all your hardware
3DMark is the only benchmark that offers a range of tests for different classes of hardware:
- Fire Strike, for high performance gaming PCs (DirectX 11, feature level 11)
- Sky Diver, for gaming laptops and mid-range PCs (DirectX 11, feature level 11)
- Cloud Gate, for notebooks and typical home PCs (DirectX 11 feature level 10)
- Ice Storm, for tablets and entry level PCs (DirectX 11 feature level 9)
With 3DMark, you can benchmark the full performance range of modern DirectX 11 graphics hardware. Where Fire Strike is like a modern game on ultra high settings, Sky Diver is closer to a DirectX 11 game played on normal settings. This makes Sky Diver the best choice for benchmarking entry level to mid-range systems and Fire Strike the perfect benchmark for high performance gaming PCs.
See 3DMark Sky Diver in full at Computex
3DMark Sky Diver will be on display on the ASUS, MSI, GIGABYTE, Galaxy, Inno3D, and G-Skill booths at Computex, June 3-7.
S.Y. Shian, ASUS Vice President & General Manager of Notebook Business Unit said,
"We are proud to partner with Futuremark to show 3DMark Sky Diver at Computex. Sky Diver helps PC gamers choose systems that offer great performance and great value. We invite everyone to visit our stand to experience 3DMark Sky Diver on a range of new ASUS products."
Sky Diver will be released as an update for all editions of 3DMark, including the free 3DMark Basic Edition.
Subject: Graphics Cards | May 28, 2014 - 11:19 AM | Ryan Shrout
Tagged: titan z, nvidia, gtx, geforce
Though delayed by a month, today marks the official release of NVIDIA's Titan Z graphics card, the dual GK110 beast with the $3000 price tag. The massive card was shown for the first time in March at NVIDIA's GPU Technology Conference and our own Tim Verry was on the grounds to get the information.
The details remain the same:
Specifically, the GTX TITAN Z is a triple slot graphics card that marries two full GK110 (big Kepler) GPUs for a total of 5,760 CUDA cores, 448 TMUs, and 96 ROPs with 12GB of GDDR5 memory on a 384-bit bus (6GB on a 384-bit bus per GPU). For the truly adventurous, it appears possible to SLI two GTX Titan Z cards using the single SLI connector. Display outputs include two DVI, one HDMI, and one DisplayPort connector.
The difference now of course is that all the clock speeds and pricing are official.
A base clock speed of 705 MHz with a Boost rate of 876 MHz places it well behind the individual GPU performance of a GeForce GTX 780 Ti or GTX Titan Black (rated at 889/980 MHz). The memory clock speed remains the same at 7.0 Gbps and you are still getting a massive 6GB of memory per GPU.
Maybe most interesting with the release of the GeForce GTX Titan Z is that NVIDIA seems to have completely fixated on non-DIY consumers with the card. We did not receive a sample of the Titan Z (nor did we get one of the Titan Black) and when I inquired as to why, NVIDIA PR stated that they were "only going to CUDA developers and system builders."
I think it is more than likely that after the release of AMD's Radeon R9 295X2 dual GPU graphics card on April 8th, with a price tag of $1500 (half of the Titan Z), the target audience was redirected. NVIDIA already had its eye on the professional markets that weren't willing to dive into the Quadro/Tesla lines (CUDA developers will likely drop $3k at the drop of a hat to get this kind of performance density). But a side benefit of creating the best flagship gaming graphics card on the planet was probably part of the story - and promptly taken away by AMD.
I still believe the Titan Z will be an impressive graphics card to behold both in terms of look and style and in terms of performance. But it would take the BIGGEST NVIDIA fans to be able to pass up buying a pair of Radeon R9 295X2 cards for a single GeForce GTX Titan Z. At least that is our assumption until we can test one for ourselves.
I'm still working to get my hands on one of these for some testing as I think the ultra high end graphics card coverage we offer is incomplete without it.
Several of NVIDIA's partners are going to be offering the Titan Z including EVGA, ASUS, MSI and Zotac. Maybe the most intersting though is EVGA's water cooled option!
So, what do you think? Anyone lining up for a Titan Z when they show up for sale?
Subject: General Tech, Graphics Cards | May 27, 2014 - 12:00 AM | Scott Michaud
Tagged: radeon, R9, R7, eyefinity, amd
AMD has just launched their Catalyst 14.6 Beta drivers for Windows and Linux. This driver will contain performance improvements for Watch Dogs, launching today in North America, and Murdered: Soul Suspect, which arrives next week. On Linux, the driver now supports Ubuntu 14.04 and its installation process has been upgraded for simplicity and user experience.
Unless performance improvements are more important to you, the biggest feature is the support for Eyefinity with mixed resolutions. With Catalyst 14.6, you no longer need a grid of identical monitors. One example use case, suggested by AMD, is a gamer who purchases an ultra-wide 2560x1080 monitor. They will be able to add a pair of 1080p monitors on either side to create a 6400x1080 viewing surface.
If the monitors are very mismatched, the driver will allow users to letterbox to the largest rectangle contained by every monitor, or "expand" to draw the largest possible rectangle (which will lead to some assets drawing outside of any monitor). A third mode, fill, behaves like Eyefinity currently does. I must give AMD a lot of credit for leaving the choice to the user.
Returning to performance with actual figures, AMD claims "up to" 25% increases in Watch Dogs at 1080p or 28% at 1600p, compared to the previous version. The new CrossFire profile also claims up to 99% scaling in that game, at 2560x1600 with 8x MSAA. Murdered: Soul Suspect will see "up to" 16% improvements on a single card, and "up to" 93% scaling. Each of these results were provided by AMD, which tested on Radeon R9 290X cards. If these CrossFire profiles (well, first, are indicative of actual performance, and) see 99% scaling across two cards, that is pretty remarkable.
A brief mention, AMD has also expanded their JPEG decoder to Kabini. Previously, it was available to Kaveri, as of Catalyst 14.1. This allows using the GPU to display images, with their test showing a series of images being processed in about half of the time. While not claimed by AMD, I expect that the GPU will also be more power-efficient (as the processor can go back to its idle state much quicker, despite activitating another component to do so). Ironically, the three images I used for this news post are encoded in PNG. You might find that amusing.
AMD Catalyst 14.6 Beta Drivers should be now available at their download site.
Subject: Graphics Cards | May 26, 2014 - 05:16 PM | Jeremy Hellstrom
Tagged: amd, radeon, r9 295x2, R9 290X
Through hard work or good luck you find yourself the proud owner of an R9 295X2 and a 4K display but somehow the performance just isn't quite good enough. You can't afford another X2 though there is an R9 290X in your price range but you just aren't sure if it will help your system out at all. That is where [H]ard|OCP steps in with this review where they prove that tri-fire in this configuration does indeed work. Not only does it work, it allows you to vastly increase your performance over a 295X2 or to improve the performance somewhat while raising your graphics settings to new highs. For those using 5760x1200 Eyefinity you probably already have your graphics options cranked; this upgrade will still offer you a linear increase in performance. Not bad if you have the money to invest!
"Will adding a single AMD Radeon R9 290X video card to the AMD Radeon R9 295X2 work? Will you get triple-GPU performance, ala TriFire CrossFire performance? This just might be a more financially feasible configuration for gamers versus QuadFire that provides a great gaming experience in Eyefinity and 4K resolutions."
Here are some more Graphics Card articles from around the web:
- Sapphire R9 290X Vapor-X OC @ Kitguru
- Sapphire Vapor-X R9 290X - Cooling the Savage Beast @HiTech Legion
- MSI Radeon R9 280X Gaming 6 GB @ techPowerUp
- Sapphire R9 290X Vapor-X OC 4GB Video Card Review @ Legit Reviews
- CLUB3D R9 290X RoyalAce Superoverclock @ Kitguru
- AMD Radeon R9 290 On Ubuntu 14.04 With Catalyst Can Beat Windows 8.1 @ Phoronix
- Catalyst On Ubuntu 14.04 Linux Competes Well With Windows 8.1 @ Phoronix
- High-End NVIDIA GeForce vs. AMD Radeon Linux Gaming Comparison @ Phoronix
- Windows 8.1 Still Outperforms Linux With Latest Intel GPU Drivers @ Phoronix
- Raijintek Morpheus @ techPowerUp
- PNY GTX 780 XLR8 OC Edition @ [H]ard|OCP
- Palit GTX780 Jetstream 6GB SLi (Ultra HD 4K) @ Kitguru
- Gigabyte GeForce GTX 750 @ Hardware Secrets
- ASUS GeForce GTX 780 Ti DirectCU II OC @ X-bit Labs
Subject: General Tech, Graphics Cards, Mobile | May 22, 2014 - 04:58 PM | Scott Michaud
Tagged: tegra k1, nvidia, iris pro, iris, Intel, hd 4000
The Chinese tech site, Evolife, acquired a few benchmarks for the Tegra K1. We do not know exactly where they got the system from, but we know that it has 4GB of RAM and 12 GB of storage. Of course, this is the version with four ARM Cortex-A15 cores (not the upcoming, 64-bit version based on Project Denver). On 3DMark Ice Storm Unlimited, it was capable of 25737 points, full system.
Image Credit: Evolife.cn
You might remember that our tests with an Intel Core i5-3317U (Ivy Bridge), back in September, achieved a score of 25630 on 3DMark Ice Storm. Of course, that was using the built-in Intel HD 4000 graphics, not a discrete solution, but it still kept up for gaming. This makes sense, though. Intel HD 4000 (GT2) graphics has a theoretical performance of 332.8 GFLOPs, while the Tegra K1 is rated at 364.8 GFLOPs. Earlier, we said that its theoretical performance is roughly on par with the GeForce 9600 GT, although the Tegra K1 supports newer APIs.
Of course, Intel has released better solutions with Haswell. Benchmarks show that Iris Pro is able to play Battlefield 4 on High settings, at 720p, with about 30FPS. The HD 4000 only gets about 12 FPS with the same configuration (and ~30 FPS on Low). This is not to compare Intel to NVIDIA's mobile part, but rather compare Tegra K1 to modern, mainstream laptops and desktops. It is getting fairly close, especially with the first wave of K1 tablets entering at the mid-$200 USD MSRP in China.
As a final note...
There was a time where Tim Sweeney, CEO of Epic Games, said that the difference between high-end and low-end PCs "is something like 100x". Scaling a single game between the two performance tiers would be next-to impossible. He noted that ten years earlier, that factor was more "10x".
Now, an original GeForce Titan is about 12x faster than the Tegra K1 and they support the same feature set. In other words, it is easier to develop a game for the PC and high-end tablet than it was to develop an PC game for high-end and low-end machines, back in 2008. PC Gaming is, once again, getting healthier.
Subject: Graphics Cards | May 15, 2014 - 06:16 PM | Ryan Shrout
Tagged: radeon, R9 290X, r9 290, r9 280x, r9 280, amd
Just the other day AMD sent out an email to the media to discuss the current pricing situation of the Radeon R9 series of graphics cards. This email started with the following statement.
You’ve seen many articles, discussions online about the AMD Radeon™ R9 lineup – especially chatter about pricing and availability. As we’ve talked about it before, the demand for the R9 lineup has been nothing but astonishing, and went well beyond our most optimistic expectations. That created a situation where gamers weren’t able to purchase their desired R9 graphics card.
Clearly AMD would not bring up the subject if the current situation was BAD news so guess what? All seems to be back normal (or expected) in terms of AMD Radeon R9 pricing and card availability. Take a look at the table below to get an idea of where Radeon's currently stand.
|Radeon R9 295X2||$1524||$1499|
|Radeon R9 290X||$549||$529|
|Radeon R9 290||$379||$399|
|Radeon R9 280X||$289||$299|
|Radeon R9 280||$249||$249|
|Radeon R9 270X||$199||$189|
|Radeon R9 270||$169||$179|
There is one price change that differs from the products' launch - the SEP of the Radeon R9 280 has dropped from $279 to $249. Nothing dramatic but a nice change.
Maybe most interesting is this line from the AMD email.
Now that product is available and at suggested pricing, these prices will remain stable. No more madness like you saw in Q1.
That emphasis is AMD's. I'm not quite sure how the company thinks they can keep a tight control on pricing now if it wasn't able to do so before, but more than likely, with the rush for coin mining hardware somewhat dying off, the prediction will hold true. (As a side note, there appears to be some discounts to be found on used Radeon hardware these days...)
Of course the AMD bundling promotion known as Never Settle Forever is still going strong with these new prices as well. Scott wrote up a story detailing this latest incarnation of the promotion and he and I both agree that while free is always
good great, the age of most of the titles in the program is a bit of a problem. But AMD did note in this email that they have "lined up a few brand new games to add to this promotion, and they'll [sic] be sharing more info with you in the next few weeks!"
Subject: General Tech, Graphics Cards, Processors, Mobile | May 15, 2014 - 05:02 PM | Scott Michaud
Tagged: nvidia, xaiomi, mipad, tegra k1
Tegra K1 is NVIDIA's new mobile processor and this first to implement the Kepler graphics architecture. In other words, it has all of the same graphics functionality as a desktop GPU with 364 GigaFLOPs of performance (a little faster than a GeForce 9600 GT). This is quite fast for a mobile product. For instance, that amount of graphics performance could max out Unreal Tournament 3 to 2560x1600 and run Crysis at 720p. Being Kepler, it supports OpenGL 4.4, OpenGL ES 3.1, DirectX 11 and 12, and GPU compute languages.
Xiaomi is launching their MiPad in Beijing, today, with an 8-inch 2048x1536 screen and the Tegra K1. They will be available in June (for China) starting at $240 USD for the 16GB version and going up to $270 for the 64GB version. Each version has 2GB of RAM, an 8MP rear-facing camera, and a 5MP front camera.
Now, we wait and see if any Tegra K1 devices come to North America and Europe - especially at that price point.
Subject: General Tech, Graphics Cards | May 12, 2014 - 08:00 PM | Scott Michaud
Tagged: titan z, nvidia, gtx titan z, geforce
To a crowd of press and developers at their GTC summit, NVIDIA announced the GeForce GTX Titan Z add-in board (AIB). Each of the two, fully unlocked, GK110 GPUs would each have access to 6GB of GDDR5 memory (12GB total). The card was expected to be available on May 8th but has yet to surface. As NVIDIA has yet to comment on the situation, many question whether it ever will.
And then we get what we think are leaked benchmarks (note: two pictures).
One concern about the Titan Z was its rated 8 TeraFLOPs of compute performance. This is a fairly sizable reduction from the theoretical maximum of 10.24 TeraFLOPs of two Titan Black processors and even less than two first-generation Titans (9 TeraFLOPs combined). We expected that this is due to reduced clock rates. What we did not expect is for benchmarks to show the GPUs boost way above those advertised levels, and even beyond the advertised boost clocks of the Titan Black and the 780 Ti. The card was seen pushing 1058 MHz in some sections, which leads to a theoretical compute performance of 12.2 TeraFLOPs (6.1 TeraFLOPs per GPU) in single precision. That is a lot.
These benchmarks also show that NVIDIA has a slight lead over AMD's R9 295X2 in many games, except Battlefield 4 and Sleeping Dogs (plus 3DMark and Unigine). Of course, these benchmarks measure the software reported frame rate and frame times and those may or may not be indicative of actual performance. While I would say that the Titan Z appears to have a slight performance lead over the R9 295X2, although a solid argument for an AMD performance win exists, it does so double the cost (at its expected $3000 USD price point). That is not up for debate.
So, until NVIDIA says anything, the Titan Z is in limbo. I am sure there exists CUDA developers who await its arrival. Personally, I would just get three Titan Blacks since you are going to need to manually schedule your workloads across multiple processors anyway (or 780 Tis if 32-bit arithmetic is enough precision). That is, of course, unless you cannot physically fit enough GeForce Titan Blacks in your motherboard and, as such, you require two GK110 chips per AIB (but not enough to bother writing a cluster scheduling application).
Subject: General Tech, Graphics Cards, Mobile | May 7, 2014 - 02:26 AM | Scott Michaud
Tagged: Thunderbolt 2, thunderbolt, nvidia, GeForce GTX 780 Ti
Externally-attached GPUs have been a topic for many years now. Numerous companies have tried, including AMD and Lucid, but no solution has ever been a widely known and available product. Even as interfaces increase in bandwidth and compatibility with internal buses, it has never been something that a laptop salesperson could suggest to users who want to dock into a high-performance station at home. At best, we are seeing it in weird "coin mining" racks to hang way more GPUs above a system than could physically mount on the motherboard.
Apparently that has not stopped the DIY community, according to chatter on Tech Inferno forums. While the above video does not really show the monitor, MacBook Pro, and GPU enclosure at the same time, let alone all wired together and on, it seems reasonable enough. The video claims to give the MacBook Pro (running Windows 8.1) access to a GeForce GTX 780 Ti with fairly high performance, despite the reduced bandwidth. Quite cool.
Subject: Graphics Cards | May 6, 2014 - 03:36 AM | Tim Verry
Tagged: r9 295x2, powercolor, hawaii, dual gpu, devil 13
PowerColor has been teasing a new graphics card on its Facebook page. The photos show a macro shot of the Devil 13 logo along with captions hitting at the new card being a dual GPU monster including one caption referring the upcoming Devil 13 as a "dual beast."
PowerColor's previous Devil 13 branded graphics card was the Radeon HD 7990 Devil 13 which contained two HD 7970 "Tahiti" GPUs on one PCB. Coincidentally, AMD recently launched a new dual GPU reference design based around two R9 290x "Hawaii" GPUs called the R9 295x2. It is still rumor and speculation at this point, but the timing and leaked photos seem to point squarely at the upcoming Devil 13 card being the first air cooled custom R9 295x2!
Adding credence to the rumors, leaked photos have appeared online with a PCB backplate that appears to match the backplate shown in the official teaser photo. The leaked photos show an absolutely beastly triple slot graphics card that places two GPUs in CrossFire on a single custom PCB powered by four 8-pin PCI-E power connectors and cooled by a gargantuan HSF comprised of an aluminum fin stack and multiple large diameter copper heatpipes along with three fans. The cooler and PCB are reinforced with brackets and a metal backplate to help keep the air cooler in pace and the PCB from bending.
If the rumors hold true, PowerColor will be unveiling the first air cooled dual GPU R9 295X2 graphics card which is an impressive feat of engineering! Using four 8-pin PCI-E power connectors definitely suggests that aftermarket overclocking is encouraged and supported even if PowerColor does not end up factory overclocking their dual GPU beast.
For reference, the stock AMD R9 295X2 features two full Hawaii GPUs with 5,632 stream processors clocked at up to 1018 MHz interfaced with 8GB of total GDDR5 memory over a 512-bit bus (each GPU has 4GB of memory and a 512-bit bus). AMD rates this configuration at 11.5 TFLOPS of single precision performance. The reference R9 295X2 has a 500W TDP and uses two 8-pin PCI-E power connectors.
Please excuse me while I wipe the drool off of my keyboard...
Stay tuned to PC Perspective for more details on the mysterious dual GPU Devil 13 from PowerColor!
In the meantime, check out our full review of the R9 295X2 (and the Hawaii architecture) and what happens when you put two R9 295X2s in Quad CrossFire into a single system for 4K gaming goodness!
Get notified when we go live!