All | Editorial | General Tech | Graphics Cards | Networking | Motherboards | Cases and Cooling | Processors | Chipsets | Memory | Displays | Systems | Storage | Mobile | Shows and Expos
Subject: Graphics Cards | September 21, 2016 - 05:39 PM | Scott Michaud
Tagged: amd, radeon, graphics drivers, crimson
Continuing with AMD's attempts, especially since the start of the Crimson Edition line, to release a driver alongside big game releases, the graphics vendor has published Radeon Software Crimson Edition 16.9.2. This one aligns with the Ultimate Edition SKU of Forza Horizon 3 from Microsoft Studios, which unlocks in a little over a day. Standard and Deluxe Edition users will need to wait until Tuesday, the 27th. As always, it rolls in all of the tweaks and fixes that AMD has found prior to the game's general release.
Also, AMD has fixed several issues, according to their pleasantly verbose release notes. Crimson Edition 16.9.2 should resolve crashes that occur in Multi-GPU mode with Ashes of the Singularity in DirectX 12. It should also fix things like mouse pointer corruption on RX 400 series graphics.
You can pick it up from AMD's website, for Windows 7, 8.1, and 10, both 32- and 64-bit versions.
Subject: Graphics Cards | September 20, 2016 - 03:58 PM | Scott Michaud
Tagged: microsoft, xbox, xbox one, pc gaming, nvidia, GTX 1080, gtx 1070
NVIDIA has just announced that specially marked, 10-series GPUs will be eligible for a Gears of War 4 download code. This bundle applies to GeForce GTX 1080 and GeForce GTX 1070 desktop GPUs, as well as laptops which integrate either of those two GPUs. As always, if you plan on purchasing a GPU due to this bundle, make sure that the product page for your retailer mentions the bundle.
Also, through the Xbox Play Anywhere initiative, NVIDIA claims that this code can be used to play the game on Xbox One as well. Xbox Play Anywhere allows users to purchase a game on either of Microsoft's software stores, Xbox Store or Windows Store, and it will automatically count as a purchase for the cross-platform equivalent. It also has implications for cloud saves, but that's a story for another day.
The bundle begins today, September 20th. Gears of War 4 launches on October 11th.
Subject: Graphics Cards | September 20, 2016 - 03:35 PM | Jeremy Hellstrom
Tagged: gigabyte, GTX 1080, GTX 1080 Xtreme Gaming Premium, factory overclocked, GIGABYTE Xtreme Engine, vr link
Gigabyte's GeForce GTX 1080 Xtreme Gaming comes with a nice overclock right out of the box, 1759MHz base, 1898MHz boost clock and a small bump to the VRAM frequency to 10.2GHz. At the push of a button you can add an extra 25MHz to the GPUs clocks assuming you install the bundled GIGABYTE Xtreme Engine which also allows you to manually tweak your settings. The Package part of the official name indicates that Gigabyte's Xtreme VR Link header panel is included with the card, you can install it in the front of your case to provide easy access to two HDMI connectors and two USB 3.0 ports for a VR headset.
Pop on over to [H]ard|OCP to see how much more they could get out of the card as well as the effect it had on gameplay.
"GIGABYTE’s GeForce GTX 1080 Xtreme Gaming Premium Pack is one premium package of goodness. Not only have we got one of the fastest GeForce GTX 1080 video cards, but GIGABYTE has thrown in the kitchen sink in this Premium Package with enthusiast oriented gaming as the focus."
Here are some more Graphics Card articles from around the web:
- MSI GTX 1070 Gaming Z 8 GB @ techPowerUp
- MSI GeForce GTX 1080 and GTX 1070 Gaming X 8G Review @ Neoseeker
- ASUS GTX 1080 & GTX 1070 STRIX OC Review @ Hardware Canucks
- ASUS GTX 1060 STRIX OC 6 GB @ techPowerUp
Subject: General Tech, Graphics Cards | September 15, 2016 - 02:51 PM | Jeremy Hellstrom
Tagged: RGB, msi, GTX 1080, EKWB, factory overclocked
MSI has just turned 30 and to help you join in the festivities they've released a custom GTX 1080 for purchase. It uses an EK Predator Liquid Cooling Unit, the card is fully covered by a waterblock and a radiator and fan are already attached. The card comes in a wooden box as a keepsake.
The card is still two slots high and the GPU is overclocked somewhat, the boost is 1860 MHz. In addition to the 30th Anniversary and MSI logos on the card, there are of course RGB lights which offer 16.8 million colours controlled by the MSI Gaming App.
Subject: General Tech, Graphics Cards | September 14, 2016 - 09:55 PM | Tim Verry
Tagged: rtg, radeon technologies group, Polaris, crimson, amd radeon, amd
It has now been a year since the formation of AMD’s Radeon Technologies Group and the graphics driven division has proven itself rather successful. Looking back with hindsight, AMD's new graphics division has enjoyed several wins with new products and advancements in driver support reclaiming market share from NVIDIA and new initiatives advancing VR, HDR, and open source visual effects.
Specifically, the Radeon Technologies Group, led by Raja Koduri, has managed to launch its new "Polaris" graphics architecture based on a 14nm FinFET process with the RX 400 series for consumers and the Radeon Pro Duo, Radeon Pro WX series, and Radeon Pro SSG (Solid State Graphics) for professionals. The company asl hit a milestone on FreeSync monitor design wins with a total of 101 displays launched to date.
Along with actual hardware, the graphics division has shaken up branding by rolling out new driver software under the Radeon Crimson Edition brand (with 21 driver releases since release) and dropping FirePro in favor of carrying over the Radeon name to create new Radeon Pro branding for its professional series of graphics cards. Driver support has also been enhanced on Linux and the AMDGPU-Pro driver works for RX 400 series.
Further, the Radeon Technologies Group launched its GPUOpen initiative back in December to foster the creation and advancement of free and open source visual effects and productivity code that developers are free to download, modify, and share.
Speaking of market share, AMD has managed to claw back some discrete GPU market share from a lowly 18% of GPUs in Q2 2015 to nearly 30% last quarter (Q2'16). That is a very respectable jump in just a year's time especially against NVIDIA's successful Pascal launches helped both by the price/performance of RX 400 as well as much needed focus on improving driver quality and timeliness of releases.
Where does this leave AMD and its RTG? Honestly, the graphics division is in a much better place than it was last year and it is in a decent position to survive and make a difference. There are still many unknowns and just as AMD's processor division is dependent on a successful Zen release, the graphics division will need Vega to be a hit in order for AMD to get wins on the high end and compete with NVIDIA on the flagship and performance fronts. They will further need Vega to update their professional series of cards many of which are still using the company's Fiji architecture which is not as power efficient as Pascal or future Volta (the competition).
With that said, the team had solid wins since their formation and are gearing up for the future. According to the announcement, the Radeon Technologies Group will be focusing on pushing virtual reality (VR) and HDR (high dynamic range) in gaming by working with developers, improving drivers, adding to their GPUOpen software collection, and launching new products.
From the press release:
"We’re passionate about perfecting pixels and delivering an unrivaled gaming experience for our community, and uncompromising power and creative freedom for developers and content creators. And if you think our first year was exciting, wait until you see what RTG has lined up for the future."
In the near future, Raja Koduri told Venture Beat to expect VR backpacks to be on show at CES in January and to look out for mobile Polaris graphics cards. Also, Radeon Crimson Edition may be incorporating features from recently acquired startup HiAlgo who developed software to dynamically monitor gameplay and adjust the resolution to maintain maximum frame rates and prevent overheating during long game sessions. One of their techniques called HiAlgo Switch would allow gamers to switch from full to half resolution (and back again) at the press of a hot-key button so as to keep FPS high if a gamer anticipates they are about to enter a demanding area that would normally result in low frame rates. While these techniques are not very important for desktop gaming (especially the CPU/GPU limiter to prevent overheating), all three would come in handy for mobile gamers using laptops with discrete cards or especially APUs.
I am looking forward to seeing where Raja and the RTG team go from here and what they have in store for AMD graphics.
Subject: Graphics Cards | September 9, 2016 - 03:59 AM | Scott Michaud
Tagged: nvidia, graphics drivers, linux
Unfortunately, I don't tend to notice when Linux drivers get released; it's something I want to report more frequently on. Luckily, this time, I heard about NVIDIA's 370.28 graphics drivers while they were still fresh. This one opens up overclocking (and underclocking) for GeForce 10-series GPUs, although NVIDIA (of course) mentions that this is “at the user's own risk”. It also fixes a bunch of Vulkan bugs.
Many of these fixes were in the previous, but beta-class drivers, 370.23. It, like 370.28, also includes experimental support for PRIME Synchronization. PRIME handles choosing which GPU drives a given display, which may be different from the GPU that is rendering that image. I'm not too familiar with the system, and I've heard some jokes from the Linux community over the last couple of years about its almost vaporware-like status, but I don't have any personal experience with it.
Subject: Graphics Cards | September 7, 2016 - 08:02 PM | Scott Michaud
Tagged: dirty pool, nvidia, geforce experience, geforce
Update (September 7th @ 9:34pm): It's been pointed out in our comments that the new GeForce Experience cannot be used without logging in. It supports NVIDIA, Google, and Facebook accounts.
It's been in Beta for a while, but NVIDIA has just officially launched their new GeForce Experience application. The release version is 22.214.171.124, so be sure to check for updates if you were in the beta and your settings panel shows an earlier version. Also, there's an “allow experimental features” checkbox right under the version number, too, also in the settings panel. It defaults to on for me, so you might want to take a look if you use GeForce Experience for anything professional (ex: Twitch streaming).
Anywho, the new version runs a lot better for me than the previous one. I used to have quite long load times, often literally in the minutes, with version 2. With version 3, it often pops up in less than a second, or maybe a couple of seconds at the worst.
Obviously, if you don't use GeForce Experience, then you don't really need to update. WHQL drivers can still be downloaded from their website (although installing drivers through GeForce Experience 3.0 has been fairly bug-free for me) and most of its other features can be obtained with other applications, like OBS Studio. That said, it's free and pretty good, so it's worth giving it a try.
Subject: Graphics Cards | September 7, 2016 - 03:41 PM | Jeremy Hellstrom
Tagged: GTX 1080 STRIX GAMING, asus, GTX 1080, aerogel
ASUS has updated their GTX 1080 Strix with a few more features than the previous models in this family. The aesthetics are a bit different but there is more to the card than that, hidden under the front edge of the card are two four pin fan headers which allow you to connect two case fans to the card which will react according to the heat load on the GPU. The new DirectCU cooler has five copper heatpipes and the PCB has eight-plus-two power-phases. There are two models, the ROG Strix-GTX1080-A8G-Gaming and the ROG Strix-GTX1080-O8G-Gaming, with core speeds of 1670/1809MHz and 1759/1898MHz respectively. The Tech Report tested out the first of those two cards, see how it matches up to the competition here.
"Asus' graphics cards are favorites of ours at TR, so we were excited when the ROG Strix GeForce GTX 1080 landed in our labs. We put it to the test to see whether Asus gave Pascal a good set of wings."
Here are some more Graphics Card articles from around the web:
- he NVIDIA GeForce GTX 1070 Founder’s Edition @ TechARP
- MSI GTX 1060 Gaming X 3 GB @ techPowerUp
- SAPPHIRE NITRO+ Radeon RX 480 @ Modders-Inc
Subject: Graphics Cards | September 6, 2016 - 05:45 PM | Scott Michaud
Tagged: nvidia, pascal, gtx 1050, geforce
I don't know why people insist on encoding screenshots from form-based windows in JPEG. You have very little color variation outside of text, which is typically thin and high-contrast from its surroundings. JPEG's Fourier Transform will cause rippling artifacts in the background, which should be solid color, and will almost definitely have a larger file size. Please, everyone, at least check to see how big a PNG will be before encoding it as JPEG. (In case you notice that I encoded it in JPEG too, that's because re-compressing JPEG artifacts makes PNG's file-size blow up, forcing me to actually need to use JPEG.)
It also makes it a bit more difficult to tell whether a screenshot has been manipulated, because the hitches make everything look suspect. Regardless, BenchLife claims to have a leaked GPU-Z result for the GeForce GTX 1050. They claim that it will be using the GP107 die at 75W, although the screenshot claims neither of these. If true, this means that it will not be a further cut-down version of GP106, as seen in the two GTX 1060 parts, which would explain a little bit why they wanted both of them to remain in the 1060 level of branding. (Although why they didn't call the 6GB version the 1060 Ti is beyond me.)
What the screenshot does suggest, though, is that it will have 4GB of GDDR5 memory, on a 128-bit bus. It will have 768 shaders, the same as the GTX 950, although clocked about 15% higher (boost vs boost) and 15W lower, bringing it back into the range of PCIe bus power (75W). That doesn't mean that it will not have a six-pin external power connector, but that could be the case, like the 750 Ti.
This would give it about 2.1 TeraFLOPs of performance, which is on part with the GeForce GTX 660 from a few generations ago, as well as the RX 460, which is also 75W TDP.
Subject: Graphics Cards | September 6, 2016 - 02:53 AM | Scott Michaud
Tagged: radeon, firepro, amd
AMD is apparently interested in supporting open-source, professional graphics. For instance, the Blender Foundation is interviewing potential hires based on a potential deal with the CPU and graphics vendor. They have also open-sourced a bunch of technologies through their GPUOpen Initiative, such as the Radeon Rays (formerly FireRays) library.
This time, at IFA 2016, they released the Radeon ProRender, which used to be called FireRender. This is a plug-in for multiple 3D applications to render high-quality, raytraced images. The open-source, third-party renderer is currently available for 3D Studio Max, in beta for Maya, Rhinoceros, and Solidworks, and coming soon for Blender. While Cycles is pretty good, the potential for cross-pollination is interesting for the future of open 3D development.
We can't go wrong with more options.
Subject: Graphics Cards | August 31, 2016 - 07:50 PM | Scott Michaud
Tagged: nvidia, gameworks, fallout 4, pc gaming
Vault 1080, which is a terrible pun by the way, is a free mod of Fallout 4 that is developed by NVIDIA Lightspeed Studios. It is designed to show off GameWorks technologies, such as volumetric lighting and HBAO+, more heavily than Bethesda did with the base game. They claim that the content lasts more than an hour, which is pretty decent for a free expansion.
It will launch on the first day of PAX West: September 2nd.
If you're wondering why NVIDIA has a game development studio, they are mostly responsible for bringing content from the PC to their Shield devices, such as Half-Life 2 and Portal. They also created NVIDIA's VR Funhouse demo, which was also release for free to show off GameWorks (such as NVIDIA Flow and VRWorks Audio) for the HTC Vive. Basically, they develop games (and now game content) to make NVIDIA's hardware more appealing.
Subject: Graphics Cards | August 31, 2016 - 06:09 PM | Jeremy Hellstrom
Tagged: radeon 16.8.3, crimson, amd
Similar to the release yesterday from NVIDIA, AMD's Crimson 16.8.3 hotfix has been timed for release with Deus Ex: Mankind Divided and the Battlefield 1 Beta. This particular update will add Crossfire profiles for both games and also fixes an unfortunate bug from the previous release which occasionally caused a static, albeit colourful screen over top of your game. Unfortunately, the Gaming Evolved overlay launch problem still exists, as does the workaround.
If you do plan on submitting bug reports whilst trying out the new Battlefield, please do head on over and upgrade so the devs are not working on issues which are already resolved.
Subject: Graphics Cards | August 31, 2016 - 05:38 PM | Jeremy Hellstrom
Tagged: amd, radeon, open source, linux, RADV, graphics driver
As of yet, AMD has not delivered the open-source Radeon Vulkan driver originally slated to arrive early this year, instead relying on their current proprietary driver. That has not stopped a team of plucky programmers from creating RADV, utilizing the existing AMDGPU LLVM compiler back-end and Intel's work with Mesa NIR intermediate representation to pass to LLVM IR. You won't get Gallium3D support, ironically RADV is too close to the metal for that to work.
Phoronix just wrapped up testing of the new driver, looking at performance for The Talos Principal and DOTA 2, contrasting the open source driver with the closed source AMDGPU-PRO. RADV is not quite 4k ready but at lower resolutions it proves very competitive.
"With word coming out last week that the RADV open-source Vulkan driver can now render Dota 2 correctly, I've been running some tests the past few days of this RADV Vulkan driver compared to AMD's official (but currently closed-source) Vulkan driver bundled with the AMDGPU-PRO Vulkan driver."
Here are some more Graphics Card articles from around the web:
- Windows 10 vs. Linux Radeon Software Performance @ Phoronix
- PowerColor Red Devil RX 480 8GB Review @ OCC
- XFX Radeon RX 460 Double Dissipation @ [H]ard|OCP
- NVIDIA GeForce GTX 1060 Founders Edition Review @ Neoseeker
Subject: General Tech, Graphics Cards | August 30, 2016 - 12:46 PM | Jeremy Hellstrom
Tagged: nvidia, GeForce 372.70, driver
NVIDIA continues with their Game Ready driver program, releasing the GeForce 372.70 driver, hand crafted in the new world by artisanal engineers to bring enhanced support to World of Warcraft: Legion, Battlefield 1: Open Beta, Deus Ex: Mankind Divided, and Quantum Break. There is not much to see in the release notes, although you can now enjoy Deus Ex in glorious 3D vision assuming you have the monitor and glasses.
If you are testing the new Battlefield you should consider updating, one would suppose the bug reports submitted using this driver will be more beneficial to the developers than an older release. You know the drill, grab them from GeForce.com or NVIDIA.com.
Subject: Graphics Cards, Motherboards | August 29, 2016 - 01:20 AM | Scott Michaud
Tagged: pcie, PCI SIG
Last week, various outlets were reporting (incorrectly) that PCIe 4.0 would provide “at least 300W” through the slot. This would have been roughly equal to the power draw that a PCIe 3.0 GPU could provide with an extra six-pin and an extra eight-pin power connector, but do so all through the slot.
Later, the PCI-SIG contacted Tom's Hardware (and likely others) to say that this is not the case. The slot will still only provide 75W of power; any other power will still need to come from external connectors. The main advantage of the standard will be extra bandwidth, about double that of PCIe 3.0, not easing cable management or making it easier to design a graphics card (by making it harder to design a motherboard).
Subject: Graphics Cards | August 24, 2016 - 10:34 AM | Ryan Shrout
Tagged: nvidia, market share, jpr, jon peddie, amd
As reported by both Mercury Research and now by Jon Peddie Research, in a graphics add-in card market that dropped dramatically in Q2 2016 in terms of total units shipped, AMD has gained significant market share against NVIDIA.
|GPU Supplier||Market share this QTR||Market share last QTR||Market share last year|
Source: Jon Peddie Research
Last year at this time, AMD was sitting at 18% market share in terms of units sold, an absolutely dismal result compared to NVIDIA's dominating 81.9%. Over the last couple of quarters we have seen AMD gain in this space, and keeping in mind that Q2 2016 does not include sales of AMD's new Polaris-based graphics cards like the Radeon RX 480, the jump to 29.9% is a big move for the company. As a result, NVIDIA falls back to 70% market share for the quarter, which is still a significant lead over the AMD.
Numbers like that shouldn't be taken lightly - for AMD to gain 7 points of market share in a single quarter indicates a substantial shift in the market. This includes all add-in cards: budget, mainstream, enthusiast and even workstation class products. One report I am received says that NVIDIA card sales specifically dropped off in Q2, though the exact reason why isn't known, and as a kind of defacto result, AMD gained sales share.
There are several other factors to watch with this data however. First, the quarterly drop in graphics card sales was -20% in Q2 when compared to Q1. That is well above the average seasonal Q1-Q2 drop, which JPR claims to be -9.7%. Much of this sell through decrease is likely due to consumers expecting releases of both NVIDIA Pascal GPUs and AMD Polaris GPUs, stalling sales as consumers delay their purchases.
The NVIDIA GeForce GTX 1080 launched on May 17th and the GTX 1070 on May 29th. The company has made very bold claims about product sales of Pascal parts so I am honestly very surprised that the overall market would drop the way it did in Q2 and that NVIDIA would fall behind AMD as much as it has. Q3 2016 may be the defining time for both GPU vendors however as it will show the results of the work put into both new architectures and both new product lines. NVIDIA reported record profits recently so it will be interesting to see how that matches up to unit sales.
Subject: Graphics Cards | August 23, 2016 - 04:18 PM | Tim Verry
Tagged: water cooling, pascal, hybrid cooler, GTX 1080, evga
EVGA recently launched a water cooled graphics card that pairs the GTX 1080 processor with the company's FTW PCB and a closed loop (AIO) water cooler to deliver a heavily overclockable card that will set you back $730.
The GTX 1080 FTW Hybrid is interesting because the company has opted to use the same custom PCB design as its FTW cards rather than a reference board. This FTW board features improved power delivery with a 10+2 power phase, two 8-pin PCI-E power connectors, Dual BIOS, and adjustable RGB LEDs. The cooler is shrouded with backlit EVGA logos and has a fan to air cool the memory and VRMs that is reportedly quiet and uses a reverse swept blade design (like their ACX air coolers) rather than a traditional blower style fan. The graphics processor is cooled by a water loop.
The water block and pump sit on top of the GPU with tubes running out to the 120mm radiator. Luckily the fan on the radiator can be easily disconnected, allowing users to use their own fan if they wish. According to Youtuber Jayztwocents, the Precision XOC software controls the fan speed of the fan on the card itself but users can not adjust the radiator fan speed themselves. You can connect your own fan to your motherboard and control it that way, however.
Display outputs include one DVI-D, one HDMI, and three DisplayPort outputs (any four of the five can be used simultaneously).
Out of the box this 215W TDP graphics card has a factory overclock of 1721 MHz base and 1860 MHz boost. Thanks to the water cooler, the GPU stays at a frosty 42°C under load. When switched to the slave BIOS (which has a higher power limit and more aggressive fan curve), the card GPU Boosted to 2025 and hit 51°C (he managed to keep that to 44°C by swapping his own EK-Vardar fan onto the radiator). Not bad, especially considering the Founder's Edition hit 85°C on air in our testing! Unfortunately, EVGA did not touch the memory and left the 8GB of GDDR5X at the stock 10 GHz.
|GTX 1080||GTX 1080 FTW Hybrid||GTX 1080 FTW Hybrid Slave BIOS|
|Rated Clock||1607 MHz||1721 MHz||1721 MHz|
|Boost Clock||1733 MHz||1860 MHz||2025 MHz|
|Memory Clock||10000 MHz||10000 MHz||10000 MHz|
|TDP||180 watts||215 watts||? watts|
|MSRP (current)||$599 ($699 FE)||$730||$730|
The water cooler should help users hit even higher overclocks and/or maintain a consistent GPU Boost clock at much lower temperatures than on air. The GTX 1080 FTW Hybrid graphics card does come at a bit of a premium at $730 (versus $699 for Founders or ~$650+ for custom models), but if you have the room in your case for the radiator this might be a nice option! (Of course custom water cooling is more fun, but it's also more expensive, time consuming, and addictive. hehe)
What do you think about these "hybrid" graphics cards?
Subject: Graphics Cards | August 23, 2016 - 01:43 PM | Jeremy Hellstrom
Tagged: amd, nvidia, Tilt Brush, VR
[H]ard|OCP continues their foray into testing VR applications, this time moving away from games to try out the rather impressive Tilt Brush VR drawing application from Google. If you have yet to see this software in action it is rather incredible, although you do still require an artist's talent and practical skills to create true 3D masterpieces.
Artisic merit may not be [H]'s strong suite but testing how well a GPU can power VR applications certainly lies within their bailiwick. Once again they tested five NVIDIA GPUs and a pair of AMD's for dropped frames and reprojection caused by a drop in FPS.
"We are changing gears a bit with our VR Performance coverage and looking at an application that is not as GPU-intensive as those we have looked at in the recent past. Google's Tilt Brush is a virtual reality application that makes use of the HTC Vive head mounted display and its motion controllers to allow you to paint in 3D space."
Here are some more Graphics Card articles from around the web:
- PowerColor Red Devil RX 470 Overclocking @ [H]ard|OCP
- MSI GeForce GTX 1060 OC 6 GB @ techPowerUp
- ASUS STRIX GAMING GTX 1070 OC @ eTeknix
- EVGA GeForce GTX 1070 FTW GAMING ACX 3.0 @ Bjorn3d
Subject: Graphics Cards | August 18, 2016 - 07:58 PM | Scott Michaud
Tagged: amd, TrueAudio, trueaudio next
Using a GPU for audio makes a lot of sense. That said, the original TrueAudio was not really about that, and it didn't really take off. The API was only implemented in a handful of titles, and it required dedicated hardware that they have since removed from their latest architectures. It was not about using the extra horsepower of the GPU to simulate sound, although they did have ideas for “sound shaders” in the original TrueAudio.
TrueAudio Next, on the other hand, is an SDK that is part of AMD's LiquidVR package. It is based around OpenCL; specifically, it uses AMD's open-source FireRays library to trace the ways that audio can move from source to receiver, including reflections. For high-frequency audio, this is a good assumption, and that range of frequencies are more useful for positional awareness in VR, anyway.
Basically, TrueAudio Next has very little to do with the original.
Interestingly, AMD is providing an interface for TrueAudio Next to reserve compute units, but optionally (and under NDA). This allows audio processing to be unhooked from the video frame rate, provided that the CPU can keep both fed with actual game data. Since audio is typically a secondary thread, it could be ready to send sound calls at any moment. Various existing portions of asynchronous compute could help with this, but allowing developers to wholly reserve a fraction of the GPU should remove the issue entirely. That said, when I was working on a similar project in WebCL, I was looking to the integrated GPU, because it's there and it's idle, so why not? I would assume that, in actual usage, CU reservation would only be enabled if an AMD GPU is the only device installed.
Anywho, if you're interested, then be sure to check out AMD's other post on it, too.
Subject: Graphics Cards | August 18, 2016 - 02:28 PM | Sebastian Peak
Tagged: nvidia, gtx 1060 3gb, gtx 1060, graphics card, gpu, geforce, 1152 CUDA Cores
NVIDIA has officially announced the 3GB version of the GTX 1060 graphics card, and it indeed contains fewer CUDA cores than the 6GB version.
The GTX 1060 Founders Edition
The product page on NVIDIA.com now reflects the 3GB model, and board partners have begun announcing their versions. The MSRP on this 3GB version is set at $199, and availablity of partner cards is expected in the next couple of weeks. The two versions will be designated only by their memory size, and no other capacities of either card are forthcoming.
|GeForce GTX 1060 3GB||GeForce GTX 1060 6GB|
|Base Clock||1506 MHz||1506 MHz|
|Boost Clock||1708 MHz||1708 MHz|
|Memory Speed||8 Gbps||8 Gbps|
As you can see from the above table, the only specification that has changed is the CUDA core count, with base/boost clocks, memory speed and interface, and TDP identical. As to performance, NVIDIA says the 6GB version holds a 5% performance advantage over this lower-cost version, which at $199 is 20% less expensive than the previous GTX 1060 6GB.