EK is releasing new GeForce GTX FE Full-Cover water blocks

Subject: Graphics Cards | May 23, 2017 - 03:58 PM |
Tagged: ek cooling, pascal, nvidia, waterblock, GTX FE

The current series of EK Cooling waterblocks for Pascal based GPUs, up to and including the new Titan X are being replaced with a new family of coolers.  The new GTX FE water blocks will be compatible with the previous generation of backplates, so you can do a partial upgrade or keep an eye out for discounts on the previous generation.

ek.jpg

These new coolers will fit on any Founders Edition reference card, from GTX 1060's through to the Titan X, currently that count stands at 106 unique graphics cards so your card is likely to be compatible.  You can choose between four models, a plain design, one with acetal, one with nickel and one with both acetal and nickel, whichever one you choose it will still run you 109.95€/$125USD

EK-FC1080_GTX_FE_CP_PCB+Coolant.jpg

Full PR is below.

EK Water Blocks, the Slovenia-based premium computer liquid cooling gear manufacturer, is releasing several new EK-FC GeForce GTX FE water blocks that are compatible with multiple reference design Founders Edition NVIDIA® GeForce GTX 1060, 1070, 1080, 1080 Ti, Titan X Pascal and Titan Xp based graphics cards. All the water blocks feature recently introduced aesthetic terminal cover as well! FE blocks come as a replacement to current GeForce GTX 10x0 / TITAN X Series of water blocks.

All current GeForce GTX 10x0 / TITAN X Series of water blocks are going to be discontinued after the stock runs out and FE blocks come as a complete replacement. FE blocks are designed to fit all reference design Founders Edition NVIDIA GeForce GTX 1060, 1070, 1080, 1080 Ti, Titan X Pascal and Titan Xp based graphics cards. The current compatibility list rounds up a total of 106 graphics cards that are on the market, but as always, we recommend that you refer to the EK Cooling Configurator for a precise compatibility match.

The new EK-FC GeForce GTX FE water blocks are also backward compatible with all EK-FC1080 GTX Backplates, EK-FC1080 GTX Ti Backplates, and EK-FC Titan X Pascal Backplates.

Availability and pricing
These water blocks are made in Slovenia, Europe and are available for purchase through EK Webshop and Partner Reseller Network. In the table below you can see manufacturer suggested retail price (MSRP) with VAT included.

Capture.PNG

 

AMD Releases Radeon Software Crimson ReLive 17.5.2

Subject: Graphics Cards | May 20, 2017 - 07:01 AM |
Tagged: graphics drivers, amd

The second graphics driver of the month from AMD, Radeon Software Crimson ReLive 17.5.2, adds optimizations for Bethesda’s new shooter, Prey. AMD claims that it will yield up to a 4.5% performance improvement, as measured on an RX 580 (versus the same card with 17.5.1). This is over and above the up to 4.7% increase that 17.5.1 had over 17.4.4.

amd-2016-crimson-relive-logo.png

Outside of that game, 17.5.2 also addresses four issues. The first is a crash in NieR: Automata. The second is long load times in Forza Horizon 3. The third is a system hang with the RX 550 when going sleep. The fourth fixed issue is a bit more complicated; apparently, in a multi-GPU system, where monitors are attached to multiple graphics cards, the primary graphics card can appear disabled in Radeon Settings. All four are now fixed, so, if they affect you, then pick up the driver.

As always, they are available from AMD’s website.

Source: AMD

Google Daydream Standalone VR Headset Powered by Snapdragon 835

Subject: Graphics Cards, Mobile | May 17, 2017 - 02:30 PM |
Tagged: snapdragon 835, snapdragon, qualcomm, google io 2017, google, daydream

During the Google I/O keynote, Google and Qualcomm announced a partnership to create a reference design for a standalone Daydream VR headset using Snapdragon 835 to enable the ecosystem of partners to have deliverable hardware in consumers’ hands by the end of 2017. The time line is aggressive, impressively so, thanks in large part to the previous work Qualcomm had done with the Snapdragon-based VR reference design we first saw in September 2016. At the time the Qualcomm platform was powered by the Snapdragon 820. Since then, Qualcomm has updated the design to integrate the Snapdragon 835 processor and platform, improving performance and efficiency along the way.

Google has now taken the reference platform and made some modifications to integrate Daydream support and will offer it to partners to show case what a standalone, untethered VR solution can do. Even though Google Daydream has been shipping in the form of slot-in phones with a “dummy” headset, integrating the whole package into a dedicate device offers several advantages.

First, I expected the free standalone units to have better performance than the phones used as a slot-in solution. With the ability to tune the device to higher thermal limits, Qualcomm and Google will be able to ramp up the clocks on the GPU and SoC to get optimal performance. And, because there is more room for a larger battery on the headset design, there should be an advantage in battery life along with the increase in performance.

sd835vr.jpg

The Qualcomm Snapdragon 835 VR Reference Device

It is also likely that the device will have better thermal properties than those using high smartphones today. In other words, with more space, there should be more area for cooling and thus the unit shouldn’t be as warm on the consumers face.

I would assume as well that the standalone units will have improved hardware over the smartphone iterations. That means better gyros, cameras, sensors, etc. that could lead to improved capability for the hardware in this form. Better hardware, tighter and more focused integration and better software support should mean lower latency and better VR gaming across the board. Assuming everything is implemented as it should.

The only major change that Google has made to this reference platform is the move away from Qualcomm’s 6DOF technology (6 degrees of freedom, allowing you to move in real space and have all necessary tracking done on the headset itself) and to Google calls WorldSense. Based on the Google Project Tango technology, this is the one area I have questions about going forward. I have used three different Tango enabled devices thus far with long-term personal testing and can say that while the possibilities for it were astounding, the implementations had been…slow. For VR that 100% cannot be the case. I don’t yet know how different its integration is from what Qualcomm had done previously, but hopefully Google will leverage the work Qualcomm has already done with its platform.

Google is claiming that consumers will have hardware based on this reference design in 2017 but no pricing has been shared with me yet. I wouldn’t expect it to be inexpensive though – we are talking about all the hardware that goes into a flagship smartphone plus a little extra for the VR goodness. We’ll see how aggressive Google wants its partners to be and if it is willing to absorb any of the upfront costs with subsidy.

Let me know if this is the direction you hope to see VR move – away from tethered PC-based solutions and into the world of standalone units.

Source: Qualcomm

Meet the GT 1030

Subject: Graphics Cards | May 17, 2017 - 01:55 PM |
Tagged: nvidia, msi, gt 1030, gigabyte, evga. zotac

The GT 1030 quietly launched from a variety of vendors late yesterday amidst the tsunami of AMD announcements.  The low profile card is advertised as offering twice the performance of the iGPU found on Intel Core i5 processors and in many cases is passively cooled.  From the pricing of the cards available now, expect to pay around $75 to $85 for this new card.

evga.jpg

EVGA announced a giveaway of several GTX 1030s at the same time as they released the model names.  The card which is currently available retails for $75 and is clocked at 1290MHz base, 1544 MHz boost and has 384 CUDA Cores.  The 2GB of GDDR5 is clocked a hair over 6GHz and runs on a 64 bit bus providing a memory bandwidth of 48.06 GB/s.  Two of their three models offer HDMI + DVI-D out, the third has a pair of DVI-D connectors.

zt-p10300a-10l_image5_0.jpg

Zotac's offering provides slightly lower clocks, a base of 1227MHz and boost of 1468MHz however the VRAM remains unchanged at 6GHz.  It pairs HDMI 2.0b with a DVI slot and comes with a low profile bracket if needed for an SFF build.

msi.png

MSI went all out and released a half dozen models, two of which you can see above.  The GT 1030 AERO ITX 2G OC is actively cooled which allows you to reach a 1265MHz base and 1518MHz boost clock.  The passively cooled GT 1030 2GH LP OCV1 runs at the same frequency and fits in a single slot externally, however you will need to leave space inside the system as the heatsink takes up an additional slot internally.  Both are fully compatible with the Afterburner Overclocking Utility and its features such as the Predator gameplay recording tool.

gb pair.png

Last but not least are a pair from Gigabyte, the GT 1030 Low Profile 2G and Silent Low Profile 2G cards.  The the cards both offer you two modes, in OC Mode the base clock is 1252MHz and boost clock 1506MHz while in Gaming Mode you will run at 1227MHz base and 1468MHz boost. 

Source: EVGA

AMD Announces Radeon Vega Frontier Edition Graphics Cards

Subject: Graphics Cards | May 16, 2017 - 07:39 PM |
Tagged: Vega, reference, radeon, graphics card, gpu, Frontier Edition, amd

AMD has revealed their concept of a premium reference GPU for the upcoming Radeon Vega launch, with the "Frontier Edition" of the new graphics cards.

Vega FE Slide.png

"Today, AMD announced its brand-new Radeon Vega Frontier Edition, the world’s most powerful solution for machine learning and advanced visualization aimed to empower the next generation of data scientists and visualization professionals -- the digital pioneers forging new paths in their fields. Designed to handle the most demanding design, rendering, and machine intelligence workloads, this powerful new graphics card excels in:

  • Machine learning. Together with AMD’s ROCm open software platform, Radeon Vega Frontier Edition enables developers to tap into the power of Vega for machine learning algorithm development. Frontier Edition delivers more than 50 percent more performance than today’s most powerful machine learning GPUs.
  • Advanced visualization. Radon Vega Frontier Edition provides the performance required to drive increasingly large and complex models for real-time visualization, physically-based rendering and virtual reality through the design phase as well as rendering phase of product development.
  • VR workloads. Radeon Vega Frontier Edition is ideal for VR content creation supporting AMD’s LiquidVR technology to deliver the gripping content, advanced visual comfort and compatibility needed for next-generation VR experiences.
  • Revolutionized game design workflows. Radeon Vega Frontier Edition simplifies and accelerates game creation by providing a single GPU optimized for every stage of a game developer’s workflow, from asset production to playtesting and performance optimization."

Vega FE.jpg

From the image provided on the official product page it appears that there will be both liquid-cooled (the gold card in the background) and air-cooled variants of these "Frontier Edition" cards, which AMD states will arrive with 16GB of HBM2 and offer 1.5x the FP32 performance and 3x the FP16 performance of the Fury X.

From AMD:

Radeon Vega Frontier Edition

  • Compute units: 64
  • Single precision compute performance (FP32): ~13 TFLOPS
  • Half precision compute performance (FP16): ~25 TFLOPS
  • Pixel Fillrate: ~90 Gpixels/sec
  • Memory capacity: 16 GBs of High Bandwidth Cache
  • Memory bandwidth: ~480 GBs/sec

The availability of the Radeon Vega Frontier Edition was announced as "late June", so we should not have too long to wait for further details, including pricing.

Source: AMD

Inno3D Introduces a Single Slot GTX 1050 Ti Graphics Card

Subject: Graphics Cards | May 13, 2017 - 11:46 PM |
Tagged: SFF, pascal, nvidia, Inno3D, GP107

Hong Kong based Inno3D recently introduced a single slot graphics card using NVIDIA’s mid-range GTX 1050 Ti GPU. The aptly named Inno3D GeForce GTX 1050 Ti (1-Slot Edition) combines the reference clocked Pascal GPU, 4GB of GDDR5 memory, and a shrouded single fan cooler clad in red and black.

Inno3D GeForce GTX 1050 Ti 1 Slot Edition.png

Around back, the card offers three display outputs including a HDMI 2.0, DisplayPort 1.4, and DVI-D. The single slot cooler is a bit of an odd design with an thin axial fan rather than a centrifugal type that sits over a fake plastic fin array. Note that these fins do not actually cool anything, in fact the PCB of the card does not even extend out to where the fan is; presumably the fins are there primarily for aesthetics and secondarily to channel a bit of the air the fan pulls down. Air is pulled in and pushed over the actual GPU heatsink (under the shroud) and out the vent holes next to the display connectors. Air is circulated through the case and is not actually exhausted like traditional dual slot (and some single slot) designs. I am curious how the choice of fan and vents will affect cooling performance.

Overclocking is going to be limited on this card, and it comes out-of-the-box clocked at NVIDIA reference speeds of 1290 MHz base and 1392 MHz boost for the GPU’s 768 cores and 7 GT/s for the 4GB of GDDR5 memory. The card measures 211 mm (~8.3”) long and should fit in just about any case. Since it pulls all of its power from the slot, it might be a good option for those slim towers OEMs like to use these days to get a bit of gaming out of a retail PC.

Inno3D is not yet talking availability or pricing, but looking at there existing lineup I would expect a MSRP around $150.

Source: Tech Report

NVIDIA Announces Tesla V100 with Volta GPU at GTC 2017

Subject: Graphics Cards | May 10, 2017 - 01:32 PM |
Tagged: v100, tesla, nvidia, gv100, gtc 2017

During the opening keynote to NVIDIA’s GPU Technology Conference, CEO Jen-Hsun Huang formally unveiled the latest GPU architecture and the first product based on it. The Tesla V100 accelerator is based on the Volta GPU architecture and features some amazingly impressive specifications. Let’s take a look.

  Tesla V100 GTX 1080 Ti Titan X (Pascal) GTX 1080 GTX 980 Ti TITAN X GTX 980 R9 Fury X R9 Fury
GPU GV100 GP102 GP102 GP104 GM200 GM200 GM204 Fiji XT Fiji Pro
GPU Cores 5120 3584 3584 2560 2816 3072 2048 4096 3584
Base Clock - 1480 MHz 1417 MHz 1607 MHz 1000 MHz 1000 MHz 1126 MHz 1050 MHz 1000 MHz
Boost Clock 1455 MHz 1582 MHz 1480 MHz 1733 MHz 1076 MHz 1089 MHz 1216 MHz - -
Texture Units 320 224 224 160 176 192 128 256 224
ROP Units 128 (?) 88 96 64 96 96 64 64 64
Memory 16GB 11GB 12GB 8GB 6GB 12GB 4GB 4GB 4GB
Memory Clock 878 MHz (?) 11000 MHz 10000 MHz 10000 MHz 7000 MHz 7000 MHz 7000 MHz 500 MHz 500 MHz
Memory Interface 4096-bit (HBM2) 352-bit 384-bit G5X 256-bit G5X 384-bit 384-bit 256-bit 4096-bit (HBM) 4096-bit (HBM)
Memory Bandwidth 900 GB/s 484 GB/s 480 GB/s 320 GB/s 336 GB/s 336 GB/s 224 GB/s 512 GB/s 512 GB/s
TDP 300 watts 250 watts 250 watts 180 watts 250 watts 250 watts 165 watts 275 watts 275 watts
Peak Compute 15 TFLOPS 10.6 TFLOPS 10.1 TFLOPS 8.2 TFLOPS 5.63 TFLOPS 6.14 TFLOPS 4.61 TFLOPS 8.60 TFLOPS 7.20 TFLOPS
Transistor Count 21.1B 12.0B 12.0B 7.2B 8.0B 8.0B 5.2B 8.9B 8.9B
Process Tech 12nm 16nm 16nm 16nm 28nm 28nm 28nm 28nm 28nm
MSRP (current) lol $699 $1,200 $599 $649 $999 $499 $649 $549

While we are low on details today, it appears that the fundamental compute units of Volta are similar to that of Pascal. The GV100 has 80 SMs with 40 TPCs and 5120 total CUDA cores, a 42% increase over the GP100 GPU used on the Tesla P100 and 42% more than the GP102 GPU used on the GeForce GTX 1080 Ti. The structure of the GPU remains the same GP100 with the CUDA cores organized as 64 single precision (FP32) per SM and 32 double precision (FP64) per SM.

image7.png

Click to Enlarge

Interestingly, NVIDIA has already told us the clock speed of this new product as well, coming in at 1455 MHz Boost, more than 100 MHz lower than the GeForce GTX 1080 Ti and 25 MHz lower than the Tesla P100.

SXM2-VoltaChipDetails.png

Click to Enlarge

Volta adds in support for a brand new compute unit though, known as Tensor Cores. With 640 of these on the GPU die, NVIDIA directly targets the neural network and deep learning fields. If this is your first time hearing about Tensor, you should read up on its influence on the hardware markets, bringing forth an open-source software library for machine learning. Google has invested in a Tensor-specific processor already, and now NVIDIA throws its hat in the ring.

Adding Tensor Cores to Volta allows the GPU to do mass processing for deep learning, on the order of a 12x improvement over Pascal’s capabilities using CUDA cores only.

07.jpg

For users interested in standard usage models, including gaming, the GV100 GPU offers 1.5x improvement in FP32 computing, up to 15 TFLOPS of theoretical performance and 7.5 TFLOPS of FP64. Other relevant specifications include 320 texture units, a 4096-bit HBM2 memory interface and 16GB of memory on-module. NVIDIA claims a memory bandwidth of 900 GB/s which works out to 878 MHz per stack.

Maybe more impressive is the transistor count: 21.1 BILLION! NVIDIA claims that this is the largest chip you can make physically with today’s technology. Considering it is being built on TSMC's 12nm FinFET technology and has an 815 mm2 die size, I see no reason to doubt them.

03.jpg

Shipping is scheduled for Q3 for Tesla V100 – at least that is when NVIDIA is promising the DXG-1 system using the chip is promised to developers.

I know many of you are interested in the gaming implications and timelines – sorry, I don’t have an answer for you yet. I will say that the bump from 10.6 TFLOPS to 15 TFLOPS is an impressive boost! But if the server variant of Volta isn’t due until Q3 of this year, I find it hard to think NVIDIA would bring the consumer version out faster than that. And whether or not NVIDIA offers gamers the chip with non-HBM2 memory is still a question mark for me and could directly impact performance and timing.

More soon!!

Source: NVIDIA

NVIDIA Releases VRWorks Audio 1.0

Subject: Graphics Cards | May 10, 2017 - 07:02 AM |
Tagged: vrworks, nvidia, audio

GPUs are good at large bundles of related tasks, saving die area by tying several chunks of data together. This is commonly used for graphics, where screens have two-to-eight million (1080p to 4K) pixels, 3d models have thousands to millions of vertexes, and so forth. Each instruction is probably done hundreds, thousands, or millions of times, and so parallelism greatly helps with utilizing real-world matter to store and translate this data.

Audio is another area with a lot of parallelism. A second of audio has tens of thousands of sound pressure samples, but another huge advantage is that higher frequency sounds model pretty decently as rays, which can be traced. NVIDIA decided to repurpose their OptiX technology into calculating these rays. Beyond the architecture demo that you often see in global illumination demos, they also integrated it into an Unreal Tournament test map.

And now it’s been released, both as a standalone SDK and as an Unreal Engine 4.15 plug-in. I don’t know what its license specifically entails, because the source code requires logging into NVIDIA’s developer portal, but it looks like the plug-ins will be available to all users of supported engines.

Source: NVIDIA

GeForce Experience 3.6 Has Vulkan and OpenGL

Subject: Graphics Cards | May 10, 2017 - 03:53 AM |
Tagged: vulkan, ShadowPlay, opengl, nvidia, geforce experience

The latest version of GeForce Experience, 3.6, adds video capture (including screenshots and live streaming) support for OpenGL and Vulkan games. The catalog of titles support by ShadowPlay, which I’m pretty sure NVIDIA wants to call Share now, despite referring to it by its old name in the blog post, now includes No Man’s Sky, DOOM, and Microsoft’s beloved OpenGL title: Minecraft.

The rest of the update focuses on tweaking a few interface elements, including its streaming panel, its video and screenshot upload panel, and its gallery. Access to the alternative graphics APIs was the clear headline-maker, however, opening the door to several large gaming groups, and potentially even more going forward.

GeForce Experience 3.6 is available now.

Source: NVIDIA

The Gigabyte Aorus GTX 1080 Ti Xtreme Edition 11G; worth the extra $50?

Subject: Graphics Cards | May 9, 2017 - 01:16 PM |
Tagged: gigabyte, aorus, 1080 ti, Xtreme Edition 11G, factory overclocked

Gigabyte's Aorus branded GTX 1080 Ti Xtreme Edition 11G just arrived on our Hardware Leaderboard, not in small part due to this review at The Tech Report.  The card utilizes the same triple fans moving air over five copper heat pipes combined with a mess of fins and a large copper plate as the non-Ti Xtreme card we have seen previously.  That cooler allows the card to be clocked at 1632MHz base, 1746MHz Boost with memory hitting over 2.8GHz right out of the box and with Afterburner you can reach even higher.  TR's testing shows that this does have a noticeable effect on performance compared to the Founder's Edition cards.

backside-logo.jpg

"Aorus' GeForce GTX 1080 Ti Xtreme Edition 11G promises to unshackle the GP102 GPU from the constraints of a reference board. We run this card through our gauntlet of performance and noise testing to see whether it's worth the premium over Nvidia's GeForce GTX 1080 Ti Founders Edition."

Here are some more Graphics Card articles from around the web:

Graphics Cards

GTC 17: NVIDIA Demos (Professional) Multi-User VR

Subject: Graphics Cards | May 9, 2017 - 07:01 AM |
Tagged: VR, quadro, nvidia, gp102

Four Quadro P6000s installed in a single server, which looks like a 4U rack-mounted box, are shown running four HTC Vive Business Edition VR systems through virtual machines. It isn’t designed to be a shipping product, just a demo for NVIDIA’s GPU Technology Conference that was developed by their engineers, but that should get the attention of this trade show’s attendees, who are mostly enterprise-focused.

nvidia-2017-fouruservrserver.png

For context, this system has roughly equivalent GPU horsepower to four Titan Xps, albeit with twice the RAM and slightly different clocks; there’s plenty of power per headset to harness. Still, running this level of high-performance application on a virtual machine could be useful in a variety of business applications, from architectural visualization to, as NVIDIA notes, amusement parks.

Given that it’s just a proof-of-concept demo, you’ll need to build it yourself to get one. They didn’t mention using any special software, though.

Source: NVIDIA

Contest: Win an ASUS GTX 1070 8GB or GTX 1050 Ti Dual-fan OC Card!

Subject: General Tech, Graphics Cards | May 5, 2017 - 10:14 AM |
Tagged: gtx 1070, GTX 1050 Ti, giveaway, geforce, contest, asus

With spring filling us with happy thoughts, our friends at ASUS are gracing us with some hardware to giveaway - that's right, it's time for a contest!

Here's what's on the docket for our readers and fans:

asuscontest.jpg

The ASUS Dual-fan line is a great option for gamers that want to balance performance and value and are quieter, cooler, and faster than reference specs. 

ASUS-2017-Giveaway-Dual_Series-Banner-v1-640x200.jpg

How do you enter? Use the form below!

Win an ASUS GeForce® GTX 1070 8GB Dual-fan OC Edition!!

I do have to apologize - this contest is open to US and Canada (except Quebec) residents only. Sorry!

A HUGE THANKS goes to our partners at ASUS for supporting PC Perspective and our raders with this great contest! Good luck to everyone!

AMD Releases Radeon Software Crimson ReLive 17.5.1

Subject: Graphics Cards | May 5, 2017 - 12:21 AM |
Tagged: amd, graphics drivers

Aligning with the release of Prey, AMD released the first Radeon Crimson ReLive graphics driver of the month: 17.5.1. This version optimizes the aforementioned title with up to a 4.7% boost in performance, versus 17.4.4 as measured on an RX 580, according to AMD. It also adds multi-GPU support to the title, for those who have multiple AMD graphics cards.

amd-2016-crimson-relive-logo.png

A bunch of bugs were also fixed in this release, as is almost always the case. Probably the most important one, though, is the patch to their auto-updater that prevents it from failing. They also fixed a couple issues with hybrid graphics, including a crash in Civilization VI with those types of systems.

You can pick up 17.5.1 from AMD’s website.

Source: AMD

Prey The Fourth Be With You??? GeForce 382.05 Drivers

Subject: Graphics Cards | May 4, 2017 - 10:24 PM |
Tagged: nvidia, graphics drivers

If you’re cringing while reading that headline, then rest assured I felt just as dirty writing it.

nvidia-geforce.png

NVIDIA has released another graphics driver, 382.05, to align with a few new game releases: Prey, Battlezone, and Gears of War 4’s latest update. The first title is a first-person action-adventure title by Arkane Studios, which releases tomorrow. Interestingly, the game runs on CryEngine... versus their internally-developed Void engine, as seen in Dishonored 2; Unreal Engine, as they’ve used with the original Dishonored; or id Tech, which Bethesda’s parent company, ZeniMax, owns through id Software and has a bit of a name-association with franchise that this Prey rebooted.

Cross-eyed yet? Good. Let’s move on.

Fans of Gears of War 4, specifically those who have multiple NVIDIA graphics cards, might be interested in SLI support for this DirectX 12-based title. As we’ve mentioned in the past, the process of load-balancing multiple GPUs has changed going from DirectX 11 to DirectX 12. According to The Coalition, SLI support was technically available on 381.89 (and 17.4.4 for AMD CrossFire), but NVIDIA is advertising it with 382.05. I’m not sure whether it’s a timing-based push, or if they optimized the experience since 381.89, but you should probably update regardless.

The driver also adds / updates the SLI profile for Sniper: Ghost Warrior 3 and Warhammer 40,000: Dawn of War III. A bunch of bugs have been fixed, too, such as “In a multi-display configuration, the extended displays are unable to enter sleep mode.” along with a couple of black and blue screen issues.

You can get them from GeForce Experience or NVIDIA’s website.

Source: NVIDIA

Benchmarking finessed Polaris, the XFX RX 580 8GB GTS Black Edition

Subject: Graphics Cards | May 1, 2017 - 02:47 PM |
Tagged: amd, RX 580, RX 580 8GB GTS Black Edition, factory overclocked, xfx

Overclockers Club takes a look at XFX's new RX 580, inside and out as they completely removed the cooler to let you see all the parts.  The card does come overclocked right out of the box, however OCC pushed the card further, hitting 1440MHz on the GPU and 2116MHz for the memory.  That result fell short of the Powercolor Golden Sample card they tested but is still not bad, as XFX is not charging much of a premium over the reference model.  Pop by to see the full results.

7.jpg

"XFX's RX 580 8GB GTS Black Edition card is a factory overclocked card that uses a TrueClock OC of 1405MHz on the core right out of the gate with an OC+ core clock of 1425MHz possible. A total of 8GB of high speed GDDR5 memory is used to handle the textures to make 2560 x 1440 the new resolution target."

Here are some more Graphics Card articles from around the web:

Graphics Cards

NVIDIA Releases GeForce 381.89 Drivers

Subject: Graphics Cards | April 28, 2017 - 09:02 PM |
Tagged: nvidia, graphics drivers

The latest Game Ready drivers from NVIDIA, 381.89, launched a couple of days before yesterday’s release of Warhammer 40,000: Dawn of War III. These drivers were the target of optimizations for that game, as well as Heroes of the Storm 2.0, Batman: Arkham VR, Rick and Morty: Virtual Rick-ality, and Wilson’s Heart.

nvidia-geforce.png

Beyond game-specific optimizations, there isn’t a whole lot of difference between these and the previous drivers, 381.65. According to the release notes, the idle voltage has been reduced in some circumstances, a crash in Sniper Elite 3 has been resolved, and two bluescreens have been fixed. That said, there’s occasionally undocumented changes that crop up.

You can pick up these drivers from NVIDIA’s website, or through GeForce Experience.

Source: NVIDIA

AMD Releases Radeon Software Crimson ReLive 17.4.4

Subject: Graphics Cards | April 28, 2017 - 03:32 PM |
Tagged: amd, graphics drivers

The latest graphics driver from AMD, Radeon Software Crimson ReLive 17.4.4, aligns with yesterday’s release of Warhammer 40,000: Dawn of War III. They claim that, when the update is applied to an 8GB Radeon RX 580, users could see a performance gain of up-to 7% under certain conditions (when compared to 17.4.3).

amd-2016-crimson-relive-logo.png

The driver was re-released on April 27th, with the new installer no-longer adding a desktop shortcut to join the Quake Champions beta. I haven’t seen it personally, but OC3D claims that the shortcut pointed to a bit.ly link. I can see why users would be upset; AMD should have added an option in the installer that says something like, “Would you like to check out the Quake Champions beta? Yes, No, Create a Desktop Shortcut for Later” rather than just add stuff to the system. That said, a desktop shortcut is as benign as you can get, and I can also see why AMD wouldn’t think much of it.

That issue aside, the driver also fixes several bugs. One notable entry is, for users with an HDR-compatible display, Mass Effect: Andromeda will now display the correct colors under Windows 10 Creators Update. The most severe fix seems to be for RX 550 users, where the GPU would hard-lock a system after “long periods of time” since the last reboot. It sounds like those users should update to 17.4.4 as soon as convenient.

Radeon Software Crimson ReLive 17.4.4 is available at AMD’s website.

Source: AMD

Sapphire Launches New Pulse Series of Graphics Cards

Subject: Graphics Cards | April 25, 2017 - 03:11 AM |
Tagged: sapphire, RX 580, RX 550, pulse, Polaris, nitro+, GCN

Earlier this month Sapphire announced a new budget-oriented series of graphics cards it calls PULSE. The new series slides in below the premium Nitro+ series to offer cheaper graphics cards that retain many of the high-quality hardware components but lack the flashy extras on the coolers, come in at lower factory overclocks, and have fewer PCI-E power inputs which, in theory, means lower overclocking headroom. The new graphics cards series is currently made up of five Polaris-based GPUs: the Sapphire Pulse RX 580, RX 570, RX 570 ITX, and RX 550.

Sapphire Pulse Graphics Cards.jpg

According to Sapphire, Pulse graphics cards use many of the high-end components as the Nitro+ cards including Black Diamond Chokes 4, long lasting capacitors, fuse protection. And intelligent fan control. The new graphics cards have aluminum backplates, removeable Quick Connect fans with semi-passive cooling technology that allows the fans to turn off when the card is under light load. The RX 580 and RX 570 use Dual-X coolers and the RX 570 ITX and RX 550 use single fan shrouded coolers.

Compared to Nitro+, the coolers are a bit less flashy and there are no Nitro+ Glow LEDs. If you are not a fan of bling or do not have a windowed case, the Pulse cards might save you a bit of money while getting you most of the performance if Sapphire’s claims are accurate.

Speaking of performance, the Pulse branded graphics cards are factory overclocked, just not as much. The Sapphire Pulse RX 580 with its 2,304 cores comes with a boost clock of 1366 MHz, the RX 570 and RX 570 ITX come with GPU boost clocks of 1,284 MHz and 1,244 MHz respectively, and the RX 550 has a boost clock of 1,206 MHz. Memory clocks sit at 8,000 MHz for the RX 580 and 7,000 MHz for the remaining Pulse cards (RX 570, RX 570 ITX, and RX 550).

Along with the introduction of its new Pulse series of graphics cards, Sapphire has entered a “strategic partnership” with motherboard manufacturer Asrock. The new graphics cards are shipping now and will be available at retailers shortly. Pricing for the RX 550 isn’t available, but prices for the other cards has appeared online as follows: Pulse RX 580 8GB for $229.99, Pulse RX 580 4GB for $199.99, Pulse RX 570 for $179.99, Pulse RX 570 ITX for $169.99.

In all, the Pulse cards appear to be about $20 cheaper than the Nitro+ variant. We will have to wait and see if those prices hold up once retailers get stock in.

Also read:

Source: Sapphire

The RX 580 on Linux, locked stock and overclock

Subject: Graphics Cards | April 24, 2017 - 06:08 PM |
Tagged: linux, RX 580, amd, overclocking, Polaris

Phoronix have had a chance to test out the refreshed Polaris RX 580 on the Linux 4.11 kernel and Mesa 17.1-devel, initially the AMDGPU-PRO 17.10 driver update was not included thanks to interesting timing.  The performance deltas are as you would expect, a slight increase in performance that is relative to the increased clock speeds, just as when run on Windows.  They also had a chance to try overclocking the new card, AMD added support for overclocking GCN 1.2 and newer cards on their proprietary Linux driver in 2016.  They managed to increase the core by 6% without running into stability issues however when they overclocked the memory, they saw serious performance decreases.  Check out the steps they tried along with the results from the overlocked GPU here.

image.php_.jpg

"Yesterday I posted the initial Radeon RX 580 Linux benchmarks while now with having more time with this "Polaris Evolved" card I've been able to try out a bit more, like the AMDGPU Linux overclocking support. Here are the ups and downs of overclocking the Radeon graphics card under Linux."

Here are some more Graphics Card articles from around the web:

Graphics Cards

Source: Phoronix

Rumor: GeForce GT 1030 Planned, Potentially with GP108

Subject: Graphics Cards | April 22, 2017 - 12:53 AM |
Tagged: pascal, nvidia, gt 1030, gp108

Expreview.com (machine-translated from Chinese) believes that NVIDIA will launch the GeForce GT 1030 to compete in the low-end. It’s difficult to tell how confident they are about this next part, due to the translation, but they believe that it will be based on a new Pascal design, GP108, rather than a further-disabled GP107 (as seen in the GTX 1050 and GTX 1050 Ti). Those parts have 640 and 768 CUDA cores, respectively, which might be where their estimate of 512 CUDA cores for GP108 comes from.

nvidia-geforce.png

As for the merits as a product, it seems a little odd to me. There is some room for it in terms of performance, sliding between the GTX 1050 and integrated graphics with a GTX 750-class part, just with higher clocks and/or lower power due to the Pascal architecture. It does seem risky, though, considering the GTX 1050 already occupies the $110 USD price point.

The post also suggests that the cards will have 1 GB and 2 GB variants.