Subject: Graphics Cards | December 21, 2015 - 01:04 PM | Jeremy Hellstrom
Tagged: GameWorks VR 1.1, nvidia, Oculus, opengl, vive
If you are blessed with the good fortune of already having a VR headset and happen to be running an NVIDIA GPU then there is a new driver you want to grab as soon as you can. The driver includes a new OpenGL extension that enables NVIDIA SLI support for OpenGL apps that display on an Oculus or Vive. NVIDIA's PR suggests you can expect your performance to improve 1.7 times, not quite doubling but certainly offering a noticeable performance improvement. The update is for both GeForce and Quadro cards.
They describe how the update will generate images in their blog post here. They imply that the changes to your code in order to benefit from this update will be minimal and it will also reduce the CPU overhead required to display the images for the right and left eye. Read on if you are interested in the developer side of this update, otherwise download your new driver and keep an eye out for application updates that enable support for SLI in VR.
Subject: Graphics Cards | December 21, 2015 - 07:25 AM | Scott Michaud
Tagged: vulkan, Mantle, Khronos, dx12, DirectX 12
The Khronos Group announced on Friday that the Vulkan API will not ship until next year. The standards body was expecting to launch it at some point in 2015. In fact, when I was first briefed on it, they specifically said that 2015 was an “under-promise and over-deliver” estimate. Vulkan is an open graphics and compute standard that was derived from AMD's Mantle. It, like OpenCL 2.1, uses the SPIR-V language for compute and shading though, which can be compiled from subsets of a variety of languages.
I know that most people will be quick to blame The Khronos Group for this, because industry bodies moving slowly is a stereotype, but I don't think it applies. When AMD created Mantle, it bore some significant delays at all levels. Its drivers and software were held back, and the public release of its SDK was delayed out of existence. Again, it would be easy to blame AMD for this, but hold on. We now get to Microsoft. DirectX 12, which is maybe even closer to Mantle than Vulkan is due to its shading language, didn't roll out as aggressively as Microsoft expected, either. Software is still pretty much non-existent when they claimed, at GDC 2014, that about 50% of PC games would be DX12-compatible by Holiday 2015. We currently have... ... zero (excluding pre-release).
Say what you like about the three examples individually, but when all three show problems, then there might just be a few issues that took longer than expected to solve. Again, this is a completely different metaphor of translating voltages coming through a PCI Express bus into fancy graphics and GPU compute, and create all of the supporting ecosystems, too.
Speaking of ecosystems, The Khronos Group has also announced that Google has upgraded their membership to “Promoter” to get more involved with Vulkan development. Google has been sort-of hostile towards certain standards from The Khronos Group on Android in the past, such as disabling OpenCL on Nexus devices, and trying to steer developers into using Android Extension Pack and Renderscript. They seem to want to use Vulkan proper this time, which is always healthy for the API.
I guess look forward to Vulkan in 2016... hopefully early.
Subject: Graphics Cards | December 17, 2015 - 05:49 PM | Jeremy Hellstrom
Tagged: radeon, crimson, amd
That's right folks, the official AMD Radeon Software Crimson Edition 15.12 has just launched for you to install. This includes the fixes for fan speeds when you are using AMD Overdrive, your settings will stick and the fans will revert to normal after you go back to the desktop from an intense gaming session. There are multiple fixes for Star Wars Battlefront, Fallout 4 and several GUI fixes within the software itself. As always there are still a few kinks being worked out but overall it is worth popping over to AMD to grab the new driver. You should also have less issues upgrading from within Crimson after this update as well.
Subject: Graphics Cards | December 16, 2015 - 08:12 AM | Scott Michaud
Tagged: nvidia, geforce experience, geforce
A new version of GeForce Experience was published yesterday. It is classified as a beta, which I'm guessing means that they wanted to release it before the Holidays, but they didn't want to have to fix potential, post-launch issues during the break. Thus, release it as a beta so users will just roll back if something doesn't work. On the other hand, NVIDIA is suggesting that it will be a recurring theme with their new "Early Access" program.
It has a few interesting features, though. First, it has a screenshot function that connects with Imgur. Steam's F12 function is pretty good for almost any title, but there are some times that you don't want to register the game with Steam, so a second option is welcome. They also have an overlay to control your stream, rather than just an indicator icon.
They added the ability to choose the Twitch Ingest Server, which is the server that the broadcaster connects to and delivers your stream into Twitch's back-end. I haven't used ShadowPlay for a while, but you previously needed to use whatever GeForce Experience chose. If it's not the best connection (ex: across the continent) then you basically had to deal with it. OBS and OBS Studio have these features too of course. I'll be clear: NVIDIA is playing catch-up to open source software in that area.
The last feature to be mentioned might just be the most interesting, though. A while ago, we mentioned that NVIDIA wants to allow online co-op by a GameStream-like service. They now have it, and it's called GameStream Co-op. The viewer can watch, take over your input, or register as a second gamepad. It requires a GeForce GTX 650 (or 660M) or higher.
Open Source your GPU!
As part of the AMD’s recent RTG (Radeon Technologies Group) Summit in Sonoma, the company released information about a new initiative to help drive development and evolution in the world of gaming called GPUOpen. As the name implies, the idea is to use an open source mentality to drivers, libraries, SDKs and more to improve the relationship between AMD’s hardware and the gaming development ecosystem.
When the current generation of consoles was first announced, AMD was riding a wave of positive PR that it hadn’t felt in many years. Because AMD Radeon hardware was at the root of the PlayStation 4 and the Xbox One, game developers would become much more adept at programming for AMD’s GCN architecture and that would waterfall down to PC gamers. At least, that was the plan. In practice though I think you’d be hard pressed to find any analyst to put their name on a statement claiming that proclamation from AMD actually transpired. It just hasn’t happened – but that does not mean that it still can’t if all the pieces fall into place.
The issue that AMD, NVIDIA, and game developers have to work around is a divided development ecosystem. While on the console side programmers tend to have very close to the metal access on CPU and GPU hardware, that hasn’t been the case with PCs until very recently. AMD was the first to make moves in this area with the Mantle API but now we have DirectX 12, a competing low level API, that will have much wider reach than Mantle or Vulkan (what Mantle has become).
AMD also believes, as do many developers, that a “black box” development environment for tools and effects packages is having a negative effect on the PC gaming ecosystem. The black box mentality means that developers don’t have access to the source code of some packages and thus cannot tweak performance and features to their liking.
Subject: Graphics Cards | December 14, 2015 - 03:55 PM | Jeremy Hellstrom
Tagged: amd, asus, STRIX R9 380X DirectCU II OC, overclock
Out of the box the ASUS STRIX R9 380X OC has a top GPU speed of 1030MHz and memory at 5.7GHz, enough to outperform a stock GTX 960 4GB at 1440p but not enough to provide satisfactory performance at that resolution. After spending some time with the card, [H]ard|OCP determined that the best overclock they could coax out of this particular GPU was 1175MHz and 6.5GHz, so they set about testing the performance at 1440p again. To make it fair they also overclocked their STRIX GTX 960 OC 4GB to 1527MHz and 8GHz. Read the full review for the detailed results, you will see that overclocking your 380X does really increase the value you get for your money.
"We take the new ASUS STRIX R9 380X DirectCU II OC based on AMD's new Radeon R9 380X GPU and overclock this video card to its highest potential. We'll compare performance in six games, including Fallout 4, to a highly overclocked ASUS GeForce GTX 960 4GB video card and find out who dominates 1440p gaming."
Here are some more Graphics Card articles from around the web:
- Sapphire R9 390 Nitro 8GB @ Kitguru
- XFX R9 380X DD XXX OC Review @ OCC
- AMD R9 380X 4GB Graphics Card CrossFire @ eTeknix
- HIS R9 380X IceQ X2 Turbo 4GB Video Card Review @ Madshrimps
Subject: Graphics Cards | December 8, 2015 - 03:56 PM | Jeremy Hellstrom
Tagged: msi, GTX 980 Ti SEA HAWK, 980 Ti, 4k
[H]ard|OCP recently wrapped up a review of the MSI GTX 980 Ti SEA HAWK and are now revisiting the GPU, focusing on the performance at 4K resolutions. The particular card that they have tops out at 1340MHz base and a 1441MHz boost clock which results in an in-game frequency of 1567MHz. For comparison testing they have tested their overclocked card against the SEA HAWK at the factory overclock and a regular 980 Ti at 4k resolution. Read on to see if this watercooled 980 Ti is worth the premium price in the full review.
"Our second installment with the MSI GTX 980 Ti SEA HAWK focuses in on gameplay at 4K and how viable it is with the fastest overclocked GTX 980 Ti we have seen yet. We will be using a reference GeForce GTX 980 Ti to show the full spectrum of the 980 Ti’s performance capabilities and emphasize the MSI GTX 980 Ti SEA HAWK’s value."
Here are some more Graphics Card articles from around the web:
- Gigabyte GTX 980 Ti Waterforce Xtreme Gaming 6 GB @ techPowerUp
- XFX R9 Fury Pro Triple Dissipation @ Bjorn3d
- PowerColor PCS+ R9 380X Myst Edition Review @ OCC
- Sapphire Radeon R9 Nitro 380X @ Kitguru
- Radeon Settings: Crimson Edition Performance Analysis @ eTeknix
- Radeon R9 Nano Small Form Factor Overclocking @ [H]ard|OCP
What RTG has planned for 2016
Last week the Radeon Technology Group invited a handful of press and analysts to a secluded location in Sonoma, CA to discuss the future of graphics, GPUs and of course Radeon. For those of you that seem a bit confused, the RTG (Radeon Technologies Group) was spun up inside AMD to encompass all of the graphics products and IP inside the company. Though today’s story is not going to focus on the fundamental changes that RTG brings to the future of AMD, I will note, without commentary, that we saw not a single AMD logo in our presentations or in the signage present throughout the week.
Much of what I learned during the RTG Summit in Sonoma is under NDA and will likely be so for some time. We learned about the future architectures, direction and product theories that will find their way into a range of solutions available in 2016 and 2017.
What I can discuss today is a pair of features that are being updated and improved for current generation graphics cards and for Radeon GPUs coming in 2016: FreeSync and HDR displays. The former is one that readers of PC Perspective should be very familiar with while the latter will offer a new window into content coming in late 2016.
High Dynamic Range Displays: Better Pixels
In just the last couple of years we have seen a spike in resolution for mobile, desktop and notebook displays. We now regularly have 4K monitors on sale for around $500 and very good quality 4K panels going for something in the $1000 range. Couple that with the increase in market share of 21:9 panels with 3440x1440 resolutions and clearly there is a demand from consumers for a better visual experience on their PCs.
But what if the answer isn’t just more pixels, but better pixels? We already have this discussed weekly when comparing render resolutions in games of 4K at lower image quality solutions versus 2560x1440 at maximum IQ settings (for example) but the truth is that panel technology has the ability to make a dramatic change to how we view all content – games, movies, productivity – with the introduction of HDR, high dynamic range.
As the slide above demonstrates there is a wide range of luminance in the real world that our eyes can see. Sunlight crosses the 1.6 billion nits mark while basic fluorescent lighting in our homes and offices exceeds 10,000 nits. Compare to the most modern PC displays that range from 0.1 nits to 250 nits and you can already tell where the discussion is heading. Even the best LCD TVs on the market today have a range of 0.1 to 400 nits.
Subject: Graphics Cards, Processors | December 8, 2015 - 08:07 AM | Scott Michaud
Tagged: hsa, GCC, amd
Phoronix, the Linux-focused hardware website, highlighted patches for the GNU Compiler Collection (GCC) that implement HSA. This will allow newer APUs, such as AMD's Carrizo, to accelerate chunks of code (mostly loops) that have been tagged with a precompiler flag as valuable to be done on the GPU. While I have done some GPGPU development, many of the low-level specifics of HSA aren't areas that I have too much experience with.
The patches have been managed by Martin Jambor of SUSE Labs. You can see a slideshow presentation of their work on the GNU website. Even though features froze about a month ago, they are apparently hoping that this will make it into the official GCC 6 release. If so, many developers around the world will be able to target HSA-compatible hardware in the first half of 2016. Technically, anyone can do so regardless, but they would need to specifically use the unofficial branch on the GCC Subversion repository. This probably means compiling it themselves, and it might even be behind on a few features in other branches that were accepted into GCC 6.
Subject: Graphics Cards | December 6, 2015 - 11:29 PM | Scott Michaud
Tagged: gigabyte, cooler master, asetek, amd
AMD and Gigabyte have each received cease and desist letters from Asetek, regarding the Radeon Fury X and GeForce GTX 980 Water Force, respectively, for using a Cooler Master-based liquid cooling solution. The Cooler Master Seiden 120M is a self-contained block and water pump, which courts have ruled that it infringes on one of Asetek's patents. Asetek has been awarded 25.375% of Cooler Master's revenue from all affected products since January 1st, 2015.
This issue obviously affects NVIDIA less than AMD, since it applies to a single product from just one AIB partner. On AMD's side, however, it affects all Fury X products, but obviously not the air-cooled Fury and Fury Nano cards. It's also possible that future SKUs could be affected as well, especially since upcoming, top end GPUs will probably be in small packages adjacent HBM 2.0 memory. This dense form-factor lends itself well to direct cooling techniques, like closed-loop water.
Even more interesting is that we believe Asetek was expecting to get the Fury X contract. We reported on an Asetek press release that claimed they received their “Largest Ever Design Win” with an undisclosed OEM. We expected it to be the follow-up to the 290X, which we assumed was called 390X because, I mean, AMD just chose that branding, right? Then the Fury X launched and it contained a Cooler Master pump. I was confused. No other candidate for “Largest Ever Design Win” popped up from Asetek, either. I guess we were right? Question mark? The press release of Asetek's design win came out in August 2014 while Asetek won the patent case in December of that year.
Regardless, this patent war has been ongoing for several months now. If it even affects any future products, I'd hope that they'd have enough warning at this point.