Subject: Graphics Cards, Systems, Mobile | July 27, 2016 - 07:58 PM | Scott Michaud
Tagged: nvidia, Nintendo, nintendo nx, tegra, Tegra X1, tegra x2, pascal, maxwell
Okay so there's a few rumors going around, mostly from Eurogamer / DigitalFoundry, that claim the Nintendo NX is going to be powered by an NVIDIA Tegra system on a chip (SoC). DigitalFoundry, specifically, cites multiple sources who claim that their Nintendo NX development kits integrate the Tegra X1 design, as seen in the Google Pixel C. That said, the Nintendo NX release date, March 2017, does provide enough time for them to switch to NVIDIA's upcoming Pascal Tegra design, rumored to be called the Tegra X2, which uses NVIDIA's custom-designed Denver CPU cores.
Preamble aside, here's what I think about the whole situation.
First, the Tegra X1 would be quite a small jump in performance over the WiiU. The WiiU's GPU, “Latte”, has 320 shaders clocked at 550 MHz, and it was based on AMD's TeraScale 1 architecture. Because these stream processors have single-cycle multiply-add for floating point values, you can get its FLOP rating by multiplying 320 shaders, 550,000,000 cycles per second, and 2 operations per clock (one multiply and one add). This yields 352 GFLOPs. The Tegra X1 is rated at 512 GFLOPs, which is just 45% more than the previous generation.
This is a very tiny jump, unless they indeed use Pascal-based graphics. If this is the case, you will likely see a launch selection of games ported from WiiU and a few games that use whatever new feature Nintendo has. One rumor is that the console will be kind-of like the WiiU controller, with detachable controllers. If this is true, it's a bit unclear how this will affect games in a revolutionary way, but we might be missing a key bit of info that ties it all together.
As for the choice of ARM over x86... well. First, this obviously allows Nintendo to choose from a wider selection of manufacturers than AMD, Intel, and VIA, and certainly more than IBM with their previous, Power-based chips. That said, it also jives with Nintendo's interest in the mobile market. They joined The Khronos Group and I'm pretty sure they've said they are interested in Vulkan, which is becoming the high-end graphics API for Android, supported by Google and others. That said, I'm not sure how many engineers exist that specialize in ARM optimization, as most mobile platforms try to abstract this as much as possible, but this could be Nintendo's attempt to settle on a standardized instruction set, and they opted for mobile over PC (versus Sony and especially Microsoft, who want consoles to follow high-end gaming on the desktop).
Why? Well that would just be speculating on speculation about speculation. I'll stop here.
Subject: Graphics Cards | July 27, 2016 - 03:43 AM | Tim Verry
Tagged: Twin Frozr VI, Radeon RX 480, polaris 10, msi
MSi is jumping full force into custom RX 480s with its upcoming line of Radeon RX 480 Gaming series including factory overclocked Gaming X and (slightly lower end) Gaming cards in both 8GB and 4GB SKUs. All four of the new graphics cards use a custom 8 phase power design, custom PCB with Military Class 4 components, and perhaps most importantly a beefy Twin Frozr VI cooler. The overclockable cards will be available by the middle of next month.
Specifically, MSI will be launching the RX 480 GAMING X 8G and RX 480 GAMING X 4G with 8GB and 4GB of GDDR5 memory respectively. These cards will have solid metal backplates and the highest factory overclocks. Below these cards sit the RX 480 GAMING 8G and RX480 GAMING 4G with the same TWIN FROZR VI cooler but sans backplate and with lower out of the box clockspeeds. Aside from those aspects, the cards all appear to offer identical features.
The new Gaming series graphics cards feature 8-pin PCI-E power connectors and 8-phase power design on a custom PCB that should allow users to push Polaris 10 quite a bit without running into issues of overheating the VRMs. The Twin Frozr VI cooler uses a nickel plated copper base plate, three 8mm copper heatpipes, a large aluminum fin array, and two large fans that spin down while the GPU temperature is under 60°C. The heatsink results in a larger than reference card that is both wider and longer at 276mm, but the size is made up for by offering 22% better cooling performance according to MSI. Further, RGB LEDs backlight the MSI logo on the side of the card. The metal backplate on the X variants should help dissipate slightly more heat than the non X models.
All for Polaris-based graphics cards offer a single DL-DVI, two HDMI, and two DisplayPort video outputs. The inclusion of two HDMI ports rather than three DP ports is allegedly to more easily support VR users by allowing them to have an HDMI connected monitor and headset connected at the same time without using adapters.
|RX 480 Gaming X 8G||RX 480 Gaming X 4G||RX 480 Gaming 8G||RX 480 Gaming 4G||RX 480 Reference|
|GPU Clock (OC Mode)||1316 MHz||1316 MHz||1292 MHz||1292 MHz||1266 MHz|
|GPU Clock (Gaming Mode)||1303 MHz||1303 MHz||1279 MHz||1279 MHz||1266 MHz|
|GPU Clock (Silent Mode)||1266 MHz||1266 MHz||1266 MHz||1266 MHz||1266 MHz|
|Memory||8GB GDDR5||4GB GDDR5||8GB GDDR5||4GB GDDR5||8GB or 4GB GDDR5|
|Memory Clock||8100 MHz||8100 MHz||8000 MHz (?)||8000 MHz (?)||8000 MHz|
|MSRP||?||?||?||?||$249 for 8GB, $199 for 4GB|
The GAMING and GAMING X RX 480s offer two tiers of factory overclocks that users can select using MSI's software utility. The non X GAMING cards will clock up to 1279 MHz in Gaming Mode and 1292 MHz in OC Mode. In Silent Mode the card will run at the same 1266 MHz boost speed as AMD's reference design card. Meanwhile the RX 480 GAMING X cards will boost up to 1303 MHz in Gaming Mode and 1316 MHz in OC Mode. In addition, MSI is bumping up the memory clockspeeds to 8100 MHz in OC Mode which is a nice surprise! MSI's announcement is not exactly clear, but it appears that the non X versions do not have factory overlcocked memory and it remains at the reference 8000 MHz.
Pricing has not yet been announced, but the cards will reportedly be on sale worldwide by mid August.
I am looking forward to seeing how far reviewers and users are able to push Polaris 10 with the Twin Frozr cooler and 8-phase VRMs!
Subject: Graphics Cards | July 27, 2016 - 01:56 AM | Tim Verry
Tagged: solid state, radeon pro, Polaris, gpgpu, amd
UPDATE (July 27th, 1am ET): More information on the Radeon Pro SSG has surfaced since the original article. According to AnandTech, the prototype graphics card actually uses an AMD Fiji GPU. The Fiji GPU is paired onboard PCI-E based storage using the same PEX8747 bridge chip used in the Radeon Pro Duo. Storage is handled by two PCI-E 3.0 x4 M.2 slots that can accommodate up to 1TB of NAND flash storage. As I mentioned below, having the storage on board the graphics card vastly reduces latency by reducing the number of hops and not having to send requests out to the rest of the system. AMD had more numbers to share following their demo, however.
From the 8K video editing demo, the dual Samsung 950 Pro PCI-E SSDs (in RAID 0) on board the Radeon Pro SSG hit 4GB/s while scrubbing through the video. That same video source stored on a Samsung 950 Pro attached to the motherboard had throughput of only 900MB/s. In theory, reaching out to system RAM still has raw throughput advantages (with DDR4 @ 3200 MHz on a Haswell-E platform theroretically capable of 62 GB/s reads and 47 GB/s writes though that would be bottlenecked by the graphics card having to go over the PCI-E 3.0 x16 link and it's maximum of 15.754 GB/s.). Of course if you can hold it in (much smaller) GDDR5 (300+GB/s depending on clocks and memory bus width) or HBM (1TB/s) and not have to go out to any other storage tier that's ideal but not always feasible especially in the HPC world.
However, having onboard storage on the same board as the GPU only a single "hop" away vastly reduces latency and offers much more total storage space than most systems have in DDR3 or DDR4. In essence, the solid state storage on the graphics card (which developers will need to specifically code for) acts as a massive cache for streaming in assets for data sets and workloads that are highly impacted by latency. This storage is not the fastest, but is the next best thing for holding active data outside of GDDR5/x or HBM. For throughput intensive workloads reaching out to system RAM will be better Finally, reaching out to system attached storage should be the last resort as it will be the slowest and most latent. Several commentors mentioned using a PCI-E based SSD in a second slot on the motherboard accessed much like GPUs in CrossFire communicate now (DMA over the PCI-E bus) which is an interesting idea that I had not considered.
Per my understanding of the situation, I think that the on board SSG storage would still be slightly more beneficial than this setup but it would get you close (I am assuming the GPU would be able to directly interact and request data from the SSD controller and not have to rely on the system CPU to do this work but I may well be mistaken. I will have to look into this further and ask the experts heh). On the prototype Radeon Pro SSG the M.2 slots are actually able to be seen as drives by the system and OS so it is essentially acting as if there was a PCI-E adapter card in a slot on the motherboard holding those drives but that may not be the case should this product actually hit the market. I do question their choice to go with Fiji rather than Polaris, but it sounds like they built the prototype off of the Radeon Pro Duo platform so I suppose it would make sense there.
Hopefully the final versions in 2017 or beyond use at least Vega though :).
Alongside the launch of new Radeon Pro WX (workstation) series graphics cards, AMD teased an interesting new Radeon Pro product: the Radeon Pro SSG. This new professional graphics card pairs a Polaris GPU with up ot a terabyte of on board solid state storage and seeks to solve one of the biggest hurdles in GP GPU performance when dealing with extremely large datasets which is latency.
One of the core focuses of AMD's HSA (heterogeneous system architecture) is unified memory and the ability of various processors (CPU, GPU, specialized co-processors, et al) to work together efficiently by being able to access and manipulate data from the same memory pool without having to copy data bck and forth between CPU-accessible memory and GPU-accessible memory. With the Radeon Pro SSG, this idea is not fully realized (it is more of a sidestep), but it will move performance further. It does not eliminate the need to copy data to the GPU before it can work on it, but once copied the GPU will be able to work on data stored in what AMD describes as a one terabyte frame buffer. This memory will be solid state and very fast, but more importantly it will be able to get at the data with much lower latency than previous methods. AMD claims the solid state storage (likely NAND but they have not said) will link with the GPU over a dedicated PCI-E bus. I suppose that if you can't bring the GPU to the data, you bring the data to the GPU!
Considering AMD's previous memory champ – the Radeon W9100 – maxed out at 32GB of GDDR5, the teased Radeon Pro SSG with its 1TB of purportedly low latency onboard flash storage opens up a slew of new possibilities for researchers and professionals in media, medical, and scientific roles working with massive datasets for imaging, creation, and simulations! I expect that there are many professionals out there eager to get their hands on one of these cards! They will be able to as well thanks to a beta program launching shortly, so long as they have $10,000 for the hardware!
AMD gave a couple of examples in their PR on the potential benefits of its "solid state graphics" including the ability to image a patient's beating heart in real time to allow medical professionals to examine and spot issues as early as possible and using the Radeon Pro SSG to edit and scrub through 8K video in real time at 90 FPS versus 17 with current offerings. On the scientific side of things being able to load up entire models into the new graphics memory (not as low latency as GDDR5 or HBM certainly) will be a boon as will being able to get data sets as close to the GPU as possible into servers using GPU accelerated databases powering websites accessed by millions of users.
It is not exactly the HSA future I have been waiting for ever so impatiently, but it is a nice advancement and an intriguing idea that I am very curious to see how well it pans out and if developers and researchers will truly take advantage of and use to further their projects. I suspect something like this could be great for deep learning tasks as well (such as powering the "clouds" behind self driving cars perhaps).
Stay tuned to PC Perspective for more information as it develops.
This is definitely a product that I will be watching and I hope that it does well. I am curious what Nvidia's and Intel's plans are here as well! What are your thoughts on AMD's "Solid State Graphics" card? All hype or something promising?
Subject: Graphics Cards | July 26, 2016 - 12:36 AM | Tim Verry
Tagged: windforce, pascal, gigabyte, GeForce GTX 1060
In a recent press release, Gigabyte announced that it will soon be adding four new GTX 1060 graphics cards to its lineup. The new cards feature Windforce series coolers and custom PCBs. At the high end is the GTX 1060 G1 Gaming followed by the GTX 1060 Windforce OC, small form factor friendly GTX 1060 Mini ITX OC, and the budget minded GTX 1060 D5. While the company has yet to divulge pricing or availability, the cards should be out within the next month or two.
All of the upcoming cards use a custom design that uses a custom PCB and power phase setup paired with Gigabyte's dual – or in the case of the Mini ITX card – single fan Windforce air cooler. Unfortunately, exact specifications for all of the cards except the high end model are unknown including core and memory clocks. The coolers use a dual composite heatpipe that directly touches the GPU to pull heat away and is dissipated by an aluminum fin stack. The fans are 90mm on all of the cards with the dual fan models using a design that has each fan spinning alternate directions of the other. The cards feature 6GB of GDDR5 memory as well as DVI, HDMI, and DisplayPort video outputs. For example, the Mini ITX OC graphics card (which is only 17cm long) and features two DVI, one HDMI, and one DP output.
More information is available on the GTX 1060 G1 Gaming. This card is a dual slot dual fan design with a 6+1 power phase (reference is 3+1) powered by a single 8-pin power connector. The fans are shrouded and there is a metal backplate to aid in stability and cooling. Gigabyte claims that its "GPU Gauntlet" technology ensures users get heavily overclockable chips thanks to sorting and using the most promising chips.
The 16nm Pascal GPU is factory overclocked to 1847 MHz boost and 1620 MHz base clockspeeds in OC mode and 1809 MHz boost and 1594 MHz base in gaming mode. Users will be able to use the company's Xtreme Engine software to dial up the overclocks further as well as mess with the RGB LEDs. For comparison, the reference clockspeeds are 1708 MHz boost and 1506 MHz base. Gigabyte has left the 6GB of GDDR5 memory untouched at 8008 MHz.
The other cards should have similarly decent factory overclocks, but it is hard to say exactly what they will be out of the box. While I am not a big fan of the aesthetics, the Windforce coolers should let users push Pascal fairly far (for air cooling).
I would guess that the Gigabyte GTX 1060 G1 Gaming will MSRP for just above $300 while the lower end cards will be around $260 (the Mini ITX OC may be at a slight premium above that).
What do you think about Gigabyte's new cards?
Subject: General Tech, Graphics Cards | July 25, 2016 - 09:48 PM | Sebastian Peak
Tagged: siggraph 2016, Siggraph, capsaicin, amd, 3D rendering
At their Capsaicin Siggraph event tonight AMD has announced that what was previously announced as the FireRender rendering engine is being officially launched as AMD Radeon ProRender, and this is becoming open-source as part of AMD's GPUOpen initiative.
From AMD's press release:
AMD today announced its powerful physically-based rendering engine is becoming open source, giving developers access to the source code.
As part of GPUOpen, Radeon ProRender (formerly previewed as AMD FireRender) enables creators to bring ideas to life through high-performance applications and workflows enhanced by photorealistic rendering.
GPUOpen is an AMD initiative designed to assist developers in creating ground-breaking games, professional graphics applications and GPU computing applications with much greater performance and lifelike experiences, at no cost and using open development tools and software.
Unlike other renderers, Radeon ProRender can simultaneously use and balance the compute capabilities of multiple GPUs and CPUs – on the same system, at the same time – and deliver state-of-the-art GPU acceleration to produce rapid, accurate results.
Radeon ProRender plugins are available today for many popular 3D content creation applications, including Autodesk® 3ds Max®, SOLIDWORKS by Dassault Systèmes and Rhino®, with Autodesk® Maya® coming soon. Radeon ProRender works across Windows®, OS X and Linux®, and supports AMD GPUs, CPUs and APUs as well as those of other vendors.
Subject: Graphics Cards | July 25, 2016 - 09:30 PM | Sebastian Peak
Tagged: siggraph 2016, Siggraph, Radeon Pro WX Series, Radeon Pro WX 7100, Radeon Pro WX 5100, Radeon Pro WX 4100, radeon, capsaicin, amd
AMD has announced new Polaris-based professional graphics cards at Siggraph 2016 this evening, with the Radeon Pro WX 4100, WX 5100, and WX 7100 GPUs.
The AMD Radeon Pro WX 7100 GPU (Image credit: AMD)
From AMD's official press release:
AMD today unveils powerful new solutions to address modern content creation and engineering: the new Radeon Pro WX Series of professional graphics cards, which harness the award-winning Polaris architecture and is designed to deliver exceptional capabilities for the immersive computing era.
Radeon Pro solutions and the new Radeon Pro WX Series of professional graphics cards represent a fundamentally different approach for professionals rooted in a commitment to open, non-proprietary software and performant, feature-rich hardware that empowers people to create the “art of the impossible”.
The new Radeon Pro WX series graphics cards deliver on the promise of this new era of creation, are optimized for open source software, and are designed for creative professionals and those pushing the boundaries of science, technology and engineering.
The AMD Radeon Pro WX 5100 GPU (Image credit: AMD)
Radeon Pro WX Series professional graphics cards are designed to address specific demands of the modern content creation era:
- Radeon Pro WX 7100 GPU is capable of handling demanding design engineering and media and entertainment workflows and is AMD’s most affordable workstation solution for professional VR content creation.
- Radeon Pro WX 5100 GPU is the ideal solution for product development, powered by the impending game-engine revolution in design visualization.
- Radeon Pro WX 4100 GPU provides great performance in a half-height design, finally bringing mid-range application performance demanded by CAD professionals to small form factor (SFF) workstations
The AMD Radeon Pro WX 4100 GPU (Image credit: AMD)
A breakdown of the known specifications for these new GPUs was provided by AnandTech in their report on the WX Series:
Subject: Graphics Cards | July 25, 2016 - 08:49 PM | Tim Verry
Tagged: sapphire, Radeon RX 480, polaris 10, nitro+, nitro
UPDATE (July 27th, 1am ET): The 8GB overclocked Sapphire Nitro+ will MSRP for $269 while the 4GB version will be $219. For more information on Sapphire's new Polaris 10 graphics card check out our archived livestream with Sapphire's Ed Crisler!
More details on custom graphics cards based around AMD's RX 480 reference GPU are starting to trickle out now that the official shipping dates are approaching (it appears many of the cards will be available next month). Sapphire is the latest AIB to provide all the juicy details on its custom Nitro+ Radeon RX 480 card!
The Nitro+ RX 480 is a dual slot card with a Dual X cooler that features two 95mm quick connect fans, vented aluminum backplate, black shroud, and aluminum heatsink. The graphics card is powered by a single 8-pin PCI-E power connector which should be enough to allow overclocking headroom and alleviate any worries over pulling too much amperage over the PEG slot on the motherboard.
Sapphire is using high end capacitors and black diamond 4 chokes. The twin fan cooler supports "quick connect" which lets users easily pull out the fans for cleaning or replacement (which seems like a neat feature considering how dusty my PC can get (it doesn't help that my corgi loves to lay against my tower heh)). RGB LEDs illuminate the Sapphire logo and fans.
Of course, all of the LEDs can be controlled by software or a button on the back of the card to change colors in response to temperatures, fan speed, cycling through all colors, and turned off completely.
The company also uses an aluminum backplate which has a nice design to it (nice to see the only part of the card most will see getting some attention for once heh) as well as vents that allow hot air to escape. Air is pulled into the card from the two fans and pushed out the back of the card and up through the backplate. I am interested to see how much this design actually improved cooling.
Rear IO includes a single DL-DVI output along with two DisplayPort 1.4 and two HDMI 2.0b video outputs. This configuration results in a smaller air intake but also lets you hook up both a HDMI monitor and VR headset. While there are five connectors, only four may be used at the same time.
While Sapphire did not touch the memory, it did factory overclock the Polaris 10 GPU to up to 1,342 MHz boost. Compared to the reference boost clockspeed of 1,266 this is a decent jump, especially for a factory out of the box overclock. Users should be able to push the GPU further though exactly how far remains to be seen and will depend on the cooler and the quality of their specific chip.
Sapphire's Nitro+ RX 480 will reportedly be available as soon as next week in both 4GB and 8GB models. The 4GB will run $220 while the 8GB card will cost $269. If these numbers hold true, that is only a $20 premium over the reference designs which certainly seems like a great value all things considered! I am looking forward to the reviews on this slick looking card and I hope that the performance and build quality are up to snuff!
Subject: Graphics Cards | July 25, 2016 - 06:51 PM | Jeremy Hellstrom
Tagged: msi, gtx 1070, Gaming Z, Twin Frozr VI, factory overclocked
The Tech Report had a chance to see what the MSI Twin Frozr VI cooler can do to a GTX 1070, they have just wrapped up a review of the Gaming Z edition of that NVIDIA card. It comes with a respectable frequency bump when you enable OC mode, 1657 MHz base and 1860 MHz boost. When they tested it under load the GPU stayed below 70C so there should be room to push the card further. Check out the full benchmark suite in their full review.
"Nvidia's second Pascal graphics card, the GeForce GTX 1070, aims to set a new bar for graphics performance in the $379-and-up price range. We put MSI's GeForce GTX 1070 Gaming Z card through the wringer to see how a more affordable Pascal card performs."
Here are some more Graphics Card articles from around the web:
- Gigabyte GeForce GTX 1070 Xtreme Gaming @ Modders-Inc
- MSI GTX 1080 Gaming X 8G RGB SLI @ Kitguru
- NVIDIA GeForce GTX 1080 Founders Edition 8GB Graphics Card Review @ NikKTech
- MSI GTX 1060 Gaming X @ eTeknix
- MSI GTX 1060 Gaming X 6G Review @ OCC
- ASUS RX 480 STRIX OC 8 GB @ techPowerUp
Subject: Graphics Cards | July 25, 2016 - 04:48 PM | Scott Michaud
Tagged: siggraph 2016, Siggraph, quadro, nvidia
SIGGRAPH is the big, professional graphics event of the year, bringing together tens of thousands of attendees. They include engineers from Adobe, AMD, Blender, Disney (including ILM, Pixar, etc.), NVIDIA, The Khronos Group, and many, many others. Not only are new products announced, but many technologies are explained in detail, down to the specific algorithms that are used, so colleagues can advance their own research and share in kind.
But new products will indeed be announced.
The NVIDIA Quadro P6000
NVIDIA, having just launched a few Pascal GPUs to other markets, decided to announce updates to their Quadro line at the event. Two cards have been added, the Quadro P5000 and the Quadro P6000, both at the top end of the product stack. Interestingly, both use GDDR5X memory, meaning that neither will be based on the GP100 design, which is built around HBM2 memory.
The NVIDIA Quadro P5000
The lower end one, the Quadro P5000, should look somewhat familiar to our reader. Exact clocks are not specified, but the chip has 2560 CUDA cores. This is identical to the GTX 1080, but with twice the memory: 16GB of GDDR5X.
Above it sits the Quadro P6000. This chip has 3840 CUDA cores, paired with 24GB of GDDR5X. We have not seen a GPU with exactly these specifications before. It has the same number of FP32 shaders as a fully unlocked GP100 die, but it doesn't have HBM2 memory. On the other hand, the new Titan X uses GP102, combining 3584 CUDA cores with GDDR5X memory, although only 12GB of it. This means that the Quadro P6000 has 256 more (single-precision) shader units than the Titan X, but otherwise very similar specifications.
Both graphics cards have four DisplayPort 1.4 connectors, as well as a single DVI output. These five connectors can be used to drive up to four, 4K, 120Hz monitors, or four, 5K, 60Hz ones. It would be nice if all five connections could be used at once, but what can you do.
Pascal has other benefits for professional users, too. For instance, Simultaneous Multi-Projection (SMP) is used in VR applications to essentially double the GPU's geometry processing ability. NVIDIA will be pushing professional VR at SIGGRAPH this year, also launching Iray VR. This uses light fields, rendered on devices like the DGX-1, with its eight GP100 chips connected by NVLink, to provide accurately lit environments. This is particularly useful for architectural visualization.
No price is given for either of these cards, but they will launch in October of this year.
Subject: Graphics Cards | July 22, 2016 - 05:51 PM | Scott Michaud
Tagged: pascal, nvidia, graphics drivers
Turns out the Pascal-based GPUs suffered from DPC latency issues, and there's been an ongoing discussion about it for a little over a month. This is not an area that I know a lot about, but it's a system that schedules workloads by priority, which provides regular windows of time for sound and video devices to update. It can be stalled by long-running driver code, though, which could manifest as stutter, audio hitches, and other performance issues. With a 10-series GeForce device installed, users have reported that this latency increases about 10-20x, from ~20us to ~300-400us. This can increase to 1000us or more under load. (8333us is ~1 whole frame at 120FPS.)
NVIDIA has acknowledged the issue and, just yesterday, released an optional hotfix. Upon installing the driver, while it could just be psychosomatic, the system felt a lot more responsive. I ran LatencyMon (DPCLat isn't compatible with Windows 8.x or Windows 10) before and after, and the latency measurement did drop significantly. It was consistently the largest source of latency, spiking in the thousands of microseconds, before the update. After the update, it was hidden by other drivers for the first night, although today it seems to have a few spikes again. That said, Microsoft's networking driver is also spiking in the ~200-300us range, so a good portion of it might be the sad state of my current OS install. I've been meaning to do a good system wipe for a while...
Measurement taken after the hotfix, while running Spotify.
That said, my computer's a mess right now.
That said, some of the post-hotfix driver spikes are reaching ~570us (mostly when I play music on Spotify through my Blue Yeti Pro). Also, Photoshop CC 2015 started complaining about graphics acceleration issues after installing the hotfix, so only install it if you're experiencing problems. About the latency, if it's not just my machine, NVIDIA might still have some work to do.
It does feel a lot better, though.