All | Editorial | General Tech | Graphics Cards | Networking | Motherboards | Cases and Cooling | Processors | Chipsets | Memory | Displays | Systems | Storage | Mobile | Shows and Expos
Subject: Graphics Cards | April 21, 2017 - 02:59 PM | Jeremy Hellstrom
Tagged: corsair, msi, gtx 1080 ti, nvidia, factory overclocked
Corsair partnered with MSI to produce the Hydro GFX GTX 1080 Ti, which has integrated watercooling based around the Corsiar H55 AiO cooler. That cooler is used to support a factory overclock of 1,506 MHz Core, 1,620 MHz Boost and a memory frequency of 11,124MHz, though with the watercooling many will seek to find exactly how much more they can squeeze out of the silicon.
According to their own testing, the GPU barely breaches 40C under load which translates into a higher sustained boost clock. The ML120 LED PWM fan attached to the radiator can be manually set to run between 400-2,400 RPM to allow users to better control how the card operates. This release adds to Corsair's previous offering, a GTX 1080 cooled with the same H55.
FREMONT, CA – April 20th, 2017 - CORSAIR®, a world leader in enthusiast memory, PC components and high-performance gaming hardware today announced the release of the new CORSAIR Hydro GFX GTX 1080 Ti all-in-one liquid cooled graphics card. Combining the extreme gaming horsepower of the NVIDIA® GeForce® GTX 1080 Ti GPU with world-renowned CORSAIR Hydro Series liquid cooling and magnetic levitation airflow, the Hydro GFX GTX 1080 Ti delivers the phenomenal performance of an overclocked GTX 1080 Ti with all the benefits of liquid cooling. Cooler temperatures, lower noise, higher clock speeds, and easy installation combine to allow the Hydro GFX GTX 1080 Ti to offer all the performance, with none of the compromises. Developed in partnership with the expert graphics team at MSI®, the Hydro GFX GTX 1080 Ti is powered by the NVIDIA GeForce GTX 1080 Ti, the most advanced NVIDIA GeForce GPU ever made. Boasting an irresponsible level of graphics processing performance, the GTX 1080 Ti features 11GB of GDDR5X memory, 3,584 CUDA Cores and a massive 352-bit memory bus, allowing it to drive today’s most demanding games and graphics applications at not just ultra-detail settings and high frame-rates, but stunning 4K resolution; it’s the ultimate GeForce GPU for PC enthusiasts who demand nothing but the best. The Hydro GFX GTX 1080 Ti pushes that performance even further, with a factory overclocked GPU core frequency of 1,506 MHz, boost frequency of 1,620 MHz and a memory frequency of 11,124 MHz, squeezing every frame per second, polygon and pixel out of the GTX 1080 Ti GPU.
The improvements aren’t just in MHz; with the Hydro GFX GTX 1080 Ti, CORSAIR takes NVIDIA’s best and makes it better. By equipping the GPU with a CORSAIR Hydro Series H55 liquid cooler, the heat produced by the GTX 1080 Ti GPU is efficiently channelled away by a micro-fin copper base to a 120mm radiator, allowing heat to be rapidly dissipated and exhausted out of your PC rather than build up inside. The result is up to 50 percent lower GPU temperatures, which in turn allows the Hydro GFX GTX 1080 Ti to boost its GPU clock speeds higher for longer, producing up to ten percent faster performance versus a stock GTX 1080 Ti. What’s more, with quick and easy installation into most 120mm case fan mounts, the Hydro GFX GTX 1080 Ti and its radiator are easy to fit into almost any PC case.
To cool the fastest GeForce GPU ever, CORSAIR selected its most advanced 120mm cooling fan, the ML120 LED. CORSAIR ML Series fans harness magnetic levitation technology to physically suspend the fan rotor away from the fan motor when in operation. This greatly reduces friction and fan noise, allowing the Hydro GFX GTX 1080 Ti to run cool while the ML120 spins at incredibly low noise levels, even with the GPU at full load. What’s more, with 4-pin PWM fan control and a 400-2,400 RPM range, you can tweak and tune the Hydro GFX GTX 1080 Ti’s cooling to suit your system – cool and quiet, or maximum airflow for the lowest temperatures and highest overclocks.
Combining the very best of NVIDIA Pascal GPU architecture, MSI graphics card design and both CORSAIR Hydro Series liquid cooling and magnetic levitation airflow, the CORSAIR Hydro GFX GTX 1080 Ti is the GTX 1080 Ti, but cooler.
- GPU: NVIDIA GeForce GTX 1080 Ti
- CUDA Cores: 3,584
- Interface: PCI Express 3.0 x16
- Boost / Base Core Clock: 1,506 MHz / 1,620 MHz (OC Mode) 1,493 MHz / 1,607 MHz (Gaming Mode) 1,480 MHz / 1,582 MHz (Silent Mode)
- Memory Clock: 11,124 MHz (OC Mode) 11,016 MHZ (Gaming Mode) 11,016 MHz (Silent Mode)
- Memory Size: 11,264MB
- Memory Type: 11GB GDDR5X
- Memory Bus: 352-bit
- Output: 3x DisplayPort (Version 1.4), 1x HDMI (Version 2.0), 1x DL-DVI-D
- Power Connector: 1x 8-pin, x 1x 6-pin
- Power Consumption: 250W Recommended
- PSU: 600W
- SKU: CB-9060010-WW
- Dimensions: Card - 269 x 111 x 35 mm, Cooler - 151 x 120 x 52 mm Weight: Card - 1,363g, Package - 2,318g
Subject: Graphics Cards | April 21, 2017 - 02:34 PM | Jeremy Hellstrom
Tagged: msi, nvidia, GAMING X+, Twin Frozr VI, GTX 1080, factory overclocked
MSI have announced a new family of GTX 1080s to the market, dubbed the GAMING X+ series. There are four models, two with 8GB of RAM and two with 6GB, all with respectable factory overclocks. All of the cards have a Twin Frozr VI cooler with double bearing TORX fans offering quiet performance as well as the capacity to allow you to push a bit more out of the card if you so desire. They also contain the mysteriously named Premium Thermal Compound X.
The MSI Gaming App has been update with a One-click to VR app that should make switching between your head mounted display and your desktop monitors quicker and easier. It also lets you switch your RGBs between 5 different patterns for those who like their GPU illuminated. Full PR below.
As the world’s leading GAMING graphics card vendor, MSI is proud to announce a new line of graphics cards based on the award-winning GAMING X series. Loaded up with faster graphics memory, the new GAMING X+ series provide an additional boost to graphics performance for smooth gameplay. Built around NVIDIA’s GeForce® GTX 10 series GPUs, the MSI GeForce® GTX 1080 GAMING X+ 8G and GeForce® GTX 1060 GAMING X+ 6G use the full force of the TWIN FROZR VI cooler, allowing for higher core and memory clock speeds for increased performance in games. The well-known shapes of the stunning TWIN FROZR cooler are intensified by a fiery red GAMING glow piercing through the cover, while the MSI GAMING dragon RGB LED on the side can be set to any of 16.8 million colors to match your mood or build. A completely new custom PCB design using Military Class 4 components enables higher overclocking performance to push your graphics card to the max. The classy matte black solid metal backplate gives the card more structural strength and provides a nice finishing touch.
As MSI’s best thermal design to date, TWIN FROZR VI has raised the bar of Graphics Card air cooling. TORX Fan 2.0 is the enhanced version of the patented TORX Fan technology which generates 22% more air pressure for better cooling performance while further reducing noise levels. On the GeForce® GTX 1080 GAMING X+, the new fans are equipped with Double Ball Bearings to ensure lasting smooth and silent operation. Connected to the huge heatsink are 8mm copper heat pipes with a squared shape at the bottom for optimal heat transfer from the solid nickel-plated copper baseplate combined with Premium Thermal Compound X to keep the Pascal powerhouse cool.
MSI Gaming App
The MSI Gaming App allows gamers to quickly switch between OC, Gaming and Silent performance modes, depending on their needs. The latest version of MSI Gaming App features One-click to VR, which instantly optimizes your PC for the best Virtual Reality experience. It also includes host of premium features like EyeRest to improve image quality and Dragon Eye which allows you to watch a YouTube video or stream while gaming. Last but not least, the Gaming App features a LED control tab, allowing gamers to choose from 5 unique lighting modes to set the right ambience for their gaming sessions with just one click.
Prepare for VR with MSI
To fully enjoy the immersive worlds of Virtual Reality, high-performance hardware is required. MSI GeForce GTX 1060 and above graphics cards deliver perfect performance for a smooth VR experience. The MSI Gaming App now features a function to optimize the user’s PC for VR performance.
Subject: Graphics Cards | April 21, 2017 - 02:12 AM | Scott Michaud
Tagged: amd, graphics drivers
During the lull in game releases, AMD has released a new graphics driver with official, WHQL-certified support for Windows 10 Creators Update. As we’ve discussed in the past, I tend to err on the side of, “If you do a decent job at internal QA and the user can choose to skip a version or three, then rapid release is probably better than sitting around for a Microsoft certificate”. I mean, why not push out fixes as they are available if there’s no obvious downsides?
Every so often, a WHQL version needs to be certified, though, if only to be accepted into Windows Update. Note that I don’t actually know whether this specific driver will be pushed by Microsoft after an update to the Creators Update – it’s just an example of a situation where WHQL matters.
That aside, the release notes for Radeon Software Crimson ReLive 17.4.3 do not state any specific fixes or changes. The main reason for this driver is to support the Creators Update for Windows 10, as well as add support for the new Radeon RX 570 and Radeon RX 580 graphics cards.
Subject: Graphics Cards | April 18, 2017 - 04:04 PM | Jeremy Hellstrom
Tagged: RX 580, radeon, Polaris, amd, powercolor, red devil
Ryan covered the improvements over the previous Polaris based cards the RX 580 offers, a higher Rated Clock and standardizing memory frequency of all RX 580 models to 8GHz. That lead to the expected increase in performance compared the the RX 480, in a marketplace somewhat different than what the first Polaris chips arrived in. Consumers now know what NVIDIA's current generation cards provide in performance and prices have settled as much as can be expected in the volatile GPU market. Those using cards several generations old may be more receptive to an upgrade than they were with the previous generation, especially as the next large launches are some time off; we shall see if this is true in the coming months.
One particular reason to consider upgrading is VR support, something [H]ard|OCP covers in their review. The improved speeds do not provide miracles in their VR Leaderboard however they do show improvements in some games such as Serious Sam, with reprojection rates dropping markedly.
"AMD is launching the AMD Radeon RX 500 series today, and we lead with a custom retail Radeon RX 580 GPU based video card from PowerColor. We’ll take the Red Devil RX 580 Golden Sample video card through the paces and see how it compares to the competition at the same price point."
Here are some more Graphics Card articles from around the web:
- AMD's Radeon RX 580 and Radeon RX 570 @ The Tech Report
- ASUS Radeon RX 580 STRIX @ Guru of 3D
- apphire Radeon RX 570 Pulse 4 GB @ techPowerUp
- Sapphire Nitro+ RX 580 8GB Review @ Neoseeker
- PowerColor Red Devil Radeon RX 580 8GB @ eTeknix
- ASUS RX 570 STRIX Gaming OC 4GB @ Kitguru
- Sapphire RX 580 Nitro+ Limited Edition 8GB @ Kitguru
- PowerColor Radeon Red Devil RX 580 8GB Golden Sample Review @ OCC
- Unigine Superposition Is A Beautiful Way To Stress Your GPU In 2017, 17-Way Graphics Card Comparison @ Phoronix
- EVGA GeForce GTX 1070 SC2 Gaming iCX Review @ Bjorn3d
- Gigabyte Aorus GTX 1080 Ti Xtreme Gaming 11 GB @ techPowerUp
Subject: Graphics Cards | April 6, 2017 - 01:57 PM | Scott Michaud
Tagged: nvidia, graphics drivers
Lining up with yesterday’s Windows 10 Creators Update opt-in, NVIDIA releases
OS Game Ready drivers. GeForce 381.65 also includes their game-specific optimizations for the Quake Champions closed beta that you have probably seen people tweeting about over the last day or so. Also, as you would expect from a graphics card and graphics driver launching on the same day, this version adds support for the new TITAN Xp.
This driver also adds Ansel support to a pair of titles: Snake Pass and Kona. Snake Pass is a puzzle platformer with a bit of a Rare art style. Kona is a mystery title with, as NVIDIA describes it, adventure, puzzle, and survival elements, set in the fictional, northern Canada village of Atamipek Lake.
You can get the new drivers from GeForce Experience or their website.
Subject: Graphics Cards | April 6, 2017 - 01:13 PM | Scott Michaud
Tagged: titan xp, pascal, nvidia
While I realize that it’s the other way around if anything, part of me wants to believe that NVIDIA released this new graphics card, the TITAN Xp, solely to prevent people from calling last year’s Titan X “Titan XP”. Alternatively, they could be trolling everyone, but doing so with a legit product launch.
The NVIDIA TITAN Xp is, finally, a fully-unlocked GP102 for the consumer market, which was previously exclusive to the Tesla P40 and Quadro P6000 graphics cards. The extra 256 CUDA cores and slight bump in boost clocks equate to an expected 10.7% increase in boost shader capacity (12.15 TFLOPs vs 10.97 TFLOPs). Memory bandwidth, for its 12GB of GDDR5X, has also increase from 480 GB/s to 547.7 GB/s, which is a 14.1% increase.
NVIDIA's blog post also mentions that macOS drivers are coming this month.
The NVIDIA TITAN Xp is available now from NVIDIA’s website for $1200 USD. 2016’s NVIDIA Titan X is also listed at $1200, but is out of stock for some weird reason… hmm. It’s almost like they released an all-around better product at the same price point.
Subject: Graphics Cards | April 4, 2017 - 09:21 PM | Scott Michaud
Tagged: graphics drivers, amd
The first AMD Radeon driver of April isn’t aligned with a major game launch. Instead, this release seems to focus on gaming technologies in general. For VR, Radeon Software Crimson ReLive 17.4.1 adds Oculus’ Asynchronous Spacewarp (ASW) to R9 Fury, R9 390, and R9 290 graphics cards. It also adds, for Windows 10, SteamVR Asynchronous Reprojection to RX 480 and RX 470 graphics cards.
The driver also adds a couple of extra display options based on the (also just added) DP1.4 HBR3 cable standard. For now, it seems like it’s just (read: “just”) 8K 60 Hz dual-cable and 8K 30Hz single-cable. The increased bandwidth also allows for several other formats, but those have nothing to do with today’s driver.
Update: AMD released a video on the same day to advertise 8K / HDR / FreeSync 2. Embed below.
A few bugs were also fixed, most of which were general bug-fixes not associated with games. Tom Clancy’s Ghost Recon Wildlands is the one exception, which should now scale better with multiple GPUs.
AMD Radeon Software Crimson ReLive 17.4.1 is now available from AMD’s website.
Subject: Graphics Cards, Mobile | April 3, 2017 - 06:18 PM | Scott Michaud
Tagged: apple, Imagination Technologies, PowerVR
This morning, Imagination Technologies Group released a press statement announcing that Apple Inc. intends to phase out their technology in 15 to 24 months. Imagination has doubts that Apple could have circumvented every piece of intellectual property, and they have requested proof from Apple that their new solution avoids all patents, trade secrets, and so forth. According to Imagination’s statement, Apple has, thus far, not provided that proof, and they don’t believe Apple’s claims.
On the one hand, it makes sense that Apple would not divulge their own trade secrets to their current-partner, soon-competitor until it’s necessary for them to do so. On the other hand, GPUs, based on previous stories, like the Intel / NVIDIA cross-license six years ago, are still a legal minefield for new players in the industry.
So, in short, Apple says they don’t need Imagination anymore, but Imagination calls bull.
From the financial side of things, Apple is a gigantic chunk of Imagination’s revenue. For the year ending on April 30th, 2016, Apple contributed about £60.7 million GBP (~$75 million USD in today’s currency) to Imagination Technology’s revenue. Over that same period, Imagination Technology’s entire revenue was £120.0 million GBP ($149.8 million USD in today’s currency).
To see how losing essentially half of your revenue can damage a company, I’ve included a screenshot of their current stock price (via Google Finance... and I apologize for the tall shot). It must be a bit scary to do business with Apple, given how much revenue they can add and subtract on a moment’s notice. I’m reminded of the iPhone 6 sapphire glass issue, where GT Advanced Technologies took on a half-billion dollars of debt to create sapphire for Apple, only to end up rejected in the end. In that case, though, Apple agreed to absolve the company of its remaining debt after GT liquidated its equipment.
As for Apple’s new GPU? It will be interesting to see how it turns out. Apple already has their own low-level graphics API, Metal, so they might have a lot to gain, although some macOS and iOS applications use OpenGL and OpenGL ES.
We’ll find out in less than two years.
Subject: Graphics Cards | April 3, 2017 - 03:15 PM | Jeremy Hellstrom
Tagged: mad max, linux, kepler, maxwell, pascal, NVIDA, vulkan, opengl
With Vulkan support being added to Mad Max, at least in beta form, Phoronix decided to take advantage of the release to test the performance of a wide variety of NVIDIA cards on the API. They grabbed over a dozen cards encompassing three different architectures, from the GTX 680 through to the GTX 1080 Ti, so you get a very good look at the change in performance of NVIDIA on Vulkan. The results are clear, in every case Vulkan was superior to OpenGL and in many cases framerate more than doubled. Drop by for a look at what some predicted was a DOA API.
"Yesterday game porter firm Feral Interactive released a public beta of Mad Max that features a Vulkan renderer in place of its OpenGL API for graphics rendering on Linux. In addition to Radeon Vulkan numbers, I posted some NVIDIA Mad Max Linux benchmarks with both renderers. Those results were exciting on the few Pascal cards tested so I have now extended that comparison to feature a line-up of 14 NVIDIA GeForce graphics cards from Kepler, Maxwell, and Pascal families while looking at this game's OpenGL vs. Vulkan performance."
Here are some more Graphics Card articles from around the web:
- MSI GTX 1080 Ti Gaming X 11 GB @ techPowerUp
- ASUS ROG STRIX GTX 1080 Ti OC @ Kitguru
- NVIDIA GeForce GTX 1080Ti Founders Edition Review @ OCC
- The GTX 1080 Ti vs. The TITAN XP Overclocking Showdown @ BabelTechReviews
- ASUS ROG STRIX GTX 1050 Ti OC 4GB @ Kitguru
Subject: Graphics Cards | March 28, 2017 - 04:32 PM | Scott Michaud
Tagged: vulkan, DirectX 12, Futuremark, 3dmark
The latest update to 3DMark adds Vulkan support to its API Overhead test, which attempts to render as many simple objects as possible while keeping above 30 FPS. This branch of game performance allows developers to add more objects into a scene, and design these art assets in a more simple, straight-forward way. This is, now, one of the first tests that can directly compare DirectX 12 and Vulkan, which we expect to be roughly equivalent, but we couldn’t tell for sure.
While I wasn’t able to run the tests myself, Luca Rocchi of Ocaholic gave it a shot on their Core i7-5820K and GTX 980. Apparently, Vulkan was just under 10% faster than DirectX 12 in their results, reaching 22.6 million draw calls in Vulkan, but 20.6 million in DirectX 12. Again, this is one test, done by a third-party, for a single system, and a single GPU driver, on a single 3D engine, and one that is designed to stress a specific portion of the API at that; take it with a grain of salt. Still, this suggests that Vulkan can keep pace with the slightly-older DirectX 12 API, and maybe even beat it.
This update also removed Mantle support. I just thought I’d mention that.
Subject: Graphics Cards | March 21, 2017 - 07:47 PM | Scott Michaud
Tagged: windows 10, vulkan, sli, multi-gpu, crossfire
Update (March 22nd @ 3:50pm EDT): And the Khronos Group has just responded to my follow-up questions. LDA has existed since Windows Vista, at the time for assisting with SLI and Crossfire support. Its implementation has changed in Windows 10, but that's not really relevant for Vulkan's multi-GPU support. To prove this, they showed LDA referenced in a Windows 8.1 MSDN post.
Vulkan's multi-GPU extensions can be used on Windows 7 and Windows 8.x. The exact process will vary from OS to OS, but the GPU vendor can implement these extensions if they choose, and LDA mode isn't exclusive to Windows 10.
Update (March 21st @ 11:55pm EDT): I came across a Microsoft Support page that discusses issues with LDA in Windows 7, so it seems like that functionality isn't limited to WDDM 2.0 and Windows 10. (Why have a support page otherwise?) Previously, I looked up an MSDN article that had it listed as a WDDM 2.0 feature, so I figured DSOGaming's assertion that it was introduced with WDDM 2.0 was correct.
As such, LDA might not require a GPU vendor's implementation at all. It'll probably be more clear when the Khronos Group responds to my earlier request, though.
That said, we're arguing over how much a GPU vendor needs to implement; either way, it will be possible to use the multi-GPU extensions in Windows 7 and Windows 8.x if the driver supports it.
Update (March 21st @ 7:30pm EDT): The Khronos Group has just released their statement. It's still a bit unclear, and I've submit another request for clarification.
Specifically, the third statement:
If an implementation on Windows does decide to use LDA mode, it is NOT tied to Windows 10. LDA mode has been available on many versions of Windows, including Windows 7 and 8.X.
... doesn't elaborate what is required for LDA mode on Windows outside of 10. (It could be Microsoft-supported, vendor-supported, or something else entirely.) I'll update again when that information is available. For now, it seems like the table, below, should actually look something like this:
|macOS||Apple doesn't allow the Vulkan API to ship in graphics drivers.
|Linux / etc.||✓||✓||✓|
... but we will update, again, should this be inaccurate.
Update (March 20th @ 3:50pm EDT): The Khronos Group has just responded that the other posts are incorrect. They haven't yet confirmed whether this post (which separates "device groups" from the more general "multi-GPU in Vulkan") is correct, though, because they're preparing an official statement. We'll update when we have more info.
Original Post Below (March 19th @ 9:36pm EDT)
A couple of days ago, some sites have noticed a bullet point that claims Windows-based GPU drivers will need WDDM in “linked display adapter” mode for “Native multi-GPU support for NVIDIA SLI and AMD Crossfire platforms” on Vulkan. This note came from an official slide deck by the Khronos Group, which was published during the recent Game Developers Conference, GDC 2017. The concern is that “linked display adapter” mode is a part of WDDM 2.0, which is exclusive to Windows 10.
This is being interpreted as “Vulkan does not support multi-GPU under Windows 7 or 8.x”.
I reached out to the Khronos Group for clarification, but I’m fairly sure I know what this does (and doesn’t) mean. Rather than starting with a written out explanation in prose, I will summarize it into a table, below, outlining what is possible on each platform. I will then elaborate below that.
|macOS||Apple doesn't allow the Vulkan API to ship in graphics drivers.
|Linux / etc.||✓||✓||✓|
So the good news is that it’s possible for a game developer to support multi-GPU (through what DirectX 12 would call MDA) on Windows 7 and Windows 8.x; the bad news is that no-one might bother with the heavy lifting. Linked display adapters allow the developer to assume that all GPUs are roughly the same performance, have the same amount of usable memory, and can be accessed through a single driver interface. On top of these assumptions, device groups also hide some annoying and tedious work inside the graphics driver, like producing a texture on one graphics card and quickly giving it to another GPU for rendering.
Basically, if the developer will go through the trouble of supporting AMD + NVIDIA or discrete GPU + integrated GPU systems, then they can support Windows 7 / 8.x in multi-GPU as well. Otherwise? Your extra GPUs will be sitting out unless you switch to DirectX 11 or OpenGL (or you use it for video encoding or something else outside the game).
On the other hand, this limitation might pressure some developers to support unlinked multi-GPU configurations. There are some interesting possibilities, including post-processing, GPGPU tasks like AI visibility and physics, and so forth, which might be ignored in titles whose developers were seduced by the simplicity of device groups. On the whole, device groups was apparently a high-priority request by game developers, and its inclusion will lead to more multi-GPU content. Developers who can justify doing it themselves, though, now have another reason to bother re-inventing a few wheels.
Or... you could just use Linux. That works, too.
Again, we are still waiting on the Khronos Group to confirm this story. See the latest update, above.
Subject: Graphics Cards | March 15, 2017 - 03:59 PM | Jeremy Hellstrom
Tagged: nvidia, GeForce 378.78
AMD has been offering support for DX12 more effectively than NVIDIA in many titles; not enough to consistently surpass the higher end GTX cards but certainly showing improvements. NVIDIA announce that their new driver, along with optimized support for the new Tom Clancy game will also offer performance increases in DX12 games. [H]ard|OCP put the numbers referenced in the PR to the test in their recent review. The news is good for the games which were mentioned but you should not expect any gains in DX11 titles with the new driver as you can see from the benchmarks results.
"We will take the new NVIDIA GeForce 378.78 performance driver and add it to our NVIDIA Video Card Driver Performance Review graphs to see if this driver has improved performance. NVIDIA has made some very bold claims lately, so let's see if those come through as true gaming advantages."
Here are some more Graphics Card articles from around the web:
Subject: General Tech, Graphics Cards | March 10, 2017 - 11:15 AM | Ryan Shrout
Tagged: video, tom petersen, pascal, nvidia, live, gtx 1080 ti, gtx, gp102, geforce
Our review of the GeForce GTX 1080 Ti 11GB graphics card is live and ready for consumption! Make sure you check it out before this afternoon's live stream!
Did you miss our GTX 1080 Ti Live Stream? Catch the reply below!
Ready your mind and body, it’s time for another GeForce GTX live stream hosted by PC Perspective’s Ryan Shrout and NVIDIA’s Tom Petersen. The general details about the GeForce GTX 1080 Ti graphics card are already official and based on the hype train and the response on social media, there is more than a little excitement.
On hand to talk about the new graphics card will be Tom Petersen, well known in our community. While the GTX 1080 Ti will be the flagship part of our live stream we will also be diving into the world of VR performance evaluation and how the new FCAT VR tool will help reviewers and standard enthusiast see where their systems stand in producing smooth, effective virtual reality gaming. We have done quite a few awesome live steams with Tom in the past, check them out if you haven't already.
NVIDIA GeForce GTX 1080 Ti and FCAT VR Live Stream
1pm PT / 4pm ET - March 9th
Need a reminder? Join our live mailing list!
The event will take place Thursday, March 9th at 4pm ET / 1pm PT at http://www.pcper.com/live. There you’ll be able to catch the live video stream as well as use our chat room to interact with the audience, asking questions for me and Tom to answer live.
Tom has a history of being both informative and entertaining and these live streaming events are always full of fun and technical information that you can get literally nowhere else. Previous streams have produced news as well – including statements on support for Adaptive Sync, release dates for displays and first-ever demos of triple display G-Sync functionality. You never know what’s going to happen or what will be said!
This just in fellow gamers: Tom is going to be providing a GeForce GTX 1080 Ti graphics card to give away during the live stream! We won't be able to ship it until the end of next week, but one lucky viewer of the live stream will be able to get their paws on the fastest graphics card we have ever tested!! Make sure you are scheduled to be here on March 9th at 1pm PT / 4pm ET!!
Win this beauty.
If you have questions, please leave them in the comments below and we'll look through them just before the start of the live stream. Of course you'll be able to tweet us questions @pcper and we'll be keeping an eye on the IRC chat as well for more inquiries. What do you want to know and hear from Tom or I?
So join us! Set your calendar for this coming Thursday at 4pm ET / 1pm PT and be here at PC Perspective to catch it. If you are a forgetful type of person, sign up for the PC Perspective Live mailing list that we use exclusively to notify users of upcoming live streaming events including these types of specials and our regular live podcast. I promise, no spam will be had!
Subject: Graphics Cards | March 10, 2017 - 02:49 AM | Scott Michaud
Tagged: nvidia, graphics drivers
Alongside the launch of the GeForce GTX 1080 Ti, NVIDIA has released a new graphics driver that, one, obviously supports the new card and, two, also rolls in a bunch of optimizations for DirectX 12 titles. The graphics vendor already announced the initiative at last week’s GDC, but it is now released and available for public use. 378.78 is also “Game Ready” for Ghost Recon Wildlands, although that’s mostly for Ansel support; most of the optimizations for Wildlands were pushed into the previous driver.
The advertised gains vary from title to title, but they claim that Rise of the Tomb Raider at 4K will jump from 20 FPS to 27 FPS. This can be viewed as either a frame rate gain of about 33%, or it can be seen as an average frame time savings of about 12ms each and every frame. If that’s what actual end-users will see -- that’s a lot!
They also note improvements in Vulkan support, too, but without any hard, numeric assertions.
If you have a GeForce 1050 Ti notebook, then this driver is also said to fix a potential bluescreen bug that you have been facing. You can pick it up from GeForce Experience or the NVIDIA website.
Subject: Graphics Cards | March 9, 2017 - 01:53 PM | Jeremy Hellstrom
Tagged: 1080 ti, geforce, gp102, gtx 1080 ti, nvidia, pascal
As you have probably noticed from our front page, today is the day we can see how the GTX 1080 Ti performs in reviewers systems. The unfortunate news is that you can't buy one yet nor do we know when you will be able to spend the $699 it will cost to order one. We can share the performance with you, once again NVIDIA's Ti model takes the top spot out performing even the $1200 TITAN X. As for overclocking the reference model, as we have not had a chance to test any cards with third party cooler on them, [H]ard|OCP were able to increase the GPU frequency over 200MHz to 1967-1987MHz in game and push the memory to 12GHz, somewhat better than what Ryan was able to. Check out their full review here, with many more just below.
"NVIDIA is launching the fastest video card it offers for gaming today in the new $699 GeForce GTX 1080 Ti. We will take this video card and test it against the GeForce GTX 1080 and GeForce GTX TITAN X at 1440p and 4K resolutions to find out how it compares. Is it really faster than a $1200 GeForce GTX TITAN X Pascal?"
Here are some more Graphics Card articles from around the web:
- Nvidia's GeForce GTX 1080 Ti @ The Tech Report
- Nvidia GTX 1080 Ti review: The fastest graphics card, again @ Ars Technica
- NVIDIA GeForce GTX 1080 Ti Founders Edition 11 GB @ techPowerUp
- The NVIDIA GTX 1080 Ti 11GB Review @ Hardware Canucks
- Nvidia GTX 1080 Ti Founders Edition 11GB @ Kitguru
- The GTX 1080 Ti Performance Review vs. the TITAN XP & the GTX 1080 @ BabelTechReviews
- Nvidia GTX 1080 Ti CPU Showdown: i7 7700k Vs Ryzen R7 1800x Vs i7 5820k @ eTeknix
- Nvidia GeForce GTX 1080 Ti 11GB @ eTeknix
- NVIDIA GeForce GTX 1080 Ti Review: A Look At 4K & Ultrawide Gaming @ Techgage
- MSI GeForce GTX 1060 Armor OC 6 GB @ techPowerUp
Subject: Graphics Cards | March 6, 2017 - 09:08 PM | Scott Michaud
Tagged: amd, graphics drivers
Just prior to the release of Tom Clancy’s Ghost Recon Wildlands, AMD has released another graphics driver with specific optimizations. Radeon Software Crimson ReLive Edition 17.3.1 is support to provide up to a 6% performance improvement (on an RX 480) in that title. It also adds a CrossFire profile under DirectX 11. Note that there’s a known issue with 3- and 4-GPU systems, which will apparently make the game crash back to desktop on launch.
Beyond this, the new graphics driver also fixes several issues, many of which involve flickering textures, objects, or mouse pointers. It also solves an issue where installing the driver could cause a failed reboot.
If you have an AMD GPU, then you can pick up the driver from their website.
Subject: Graphics Cards | March 1, 2017 - 05:04 PM | Sebastian Peak
Tagged: video card, RX 580, RX 570, RX 560, RX 550, rx 480, rumor, report, rebrand, radeon, graphics, gpu, amd
According to a report from VideoCardz.com we can expect AMD Radeon RX 500-series graphics cards next month, with an April 4th launch of the RX 580 and RX 570, and subsequent RX 560/550 launch on April 11. The bad news? According to the report "all cards, except RX 550, are most likely rebranded from Radeon RX 400 series".
Until official confirmation on specs arrive, this is still speculative; however, if Vega is not ready for an April launch and AMD will indeed be refreshing their Radeon lineup, an R9 300-series speed bump/rebrand is not out of the realm of possibility. VideoCardz offers (unconfirmed, at this point) specs of the upcoming RX 500-series cards, with RX 400 numbers for comparison:
Chart credit: VideoCardz.com
The first graph shows the increased GPU boost clock speed of ~1340 MHz for the rumored RX 580, with the existing RX 480 clocked at 1266 MHz. Both would be Polaris 10 GPUs with otherwise identical specs. The same largely holds for the rumored specs on the RX 570, though this GPU would presumably be shipping with faster memory clocks as well. On the RX 560 side, however, the Polaris 11 powered replacement for the RX 460 might be based on the 1024-core variant we have seen from the Chinese market.
Chart credit: VideoCardz.com
No specifics on the RX 550 are yet known, which VideoCardz says "is most likely equipped with Polaris 12, a new low-end GPU". These rumors come via heise.de (German language), who state that those "hoping for Vega-card will be disappointed - the cards are intended to be rebrands with known GPUs". We will have to wait until next month to know for sure, but even if this is the case, expect faster clocks and better performance for the same money.
Subject: Graphics Cards | February 28, 2017 - 10:59 PM | Ryan Shrout
Tagged: pascal, nvidia, gtx 1080 ti, gp102, geforce
Tonight at a GDC party hosted by CEO Jen-Hsun Huang, NVIDIA announced the GeForce GTX 1080 Ti graphics card, coming next week for $699. Let’s dive right into the specifications!
|GTX 1080 Ti||Titan X (Pascal)||GTX 1080||GTX 980 Ti||TITAN X||GTX 980||R9 Fury X||R9 Fury||R9 Nano|
|GPU||GP102||GP102||GP104||GM200||GM200||GM204||Fiji XT||Fiji Pro||Fiji XT|
|Base Clock||1480 MHz||1417 MHz||1607 MHz||1000 MHz||1000 MHz||1126 MHz||1050 MHz||1000 MHz||up to 1000 MHz|
|Boost Clock||1600 MHz||1480 MHz||1733 MHz||1076 MHz||1089 MHz||1216 MHz||-||-||-|
|Memory Clock||11000 MHz||10000 MHz||10000 MHz||7000 MHz||7000 MHz||7000 MHz||500 MHz||500 MHz||500 MHz|
|Memory Interface||352-bit||384-bit G5X||256-bit G5X||384-bit||384-bit||256-bit||4096-bit (HBM)||4096-bit (HBM)||4096-bit (HBM)|
|Memory Bandwidth||484 GB/s||480 GB/s||320 GB/s||336 GB/s||336 GB/s||224 GB/s||512 GB/s||512 GB/s||512 GB/s|
|TDP||250 watts||250 watts||180 watts||250 watts||250 watts||165 watts||275 watts||275 watts||175 watts|
|Peak Compute||10.6 TFLOPS||10.1 TFLOPS||8.2 TFLOPS||5.63 TFLOPS||6.14 TFLOPS||4.61 TFLOPS||8.60 TFLOPS||7.20 TFLOPS||8.19 TFLOPS|
The GTX 1080 Ti looks a whole lot like the TITAN X launched in August of last year. Based on the 12B transistor GP102 chip, the new GTX 1080 Ti will have 3,584 CUDA core with a 1.60 GHz Boost clock. That gives it the same processor count as Titan X but with a slightly higher clock speed which should make the new GTX 1080 Ti slightly faster by at least a few percentage points and has a 4.7% edge in base clock compute capability. It has 28 SMs, 28 geometry units, 224 texture units.
Interestingly, the memory system on the GTX 1080 Ti gets adjusted – NVIDIA has disabled a single 32-bit memory controller to give the card a total of 352-bit wide bus and an odd-sounding 11GB memory capacity. The ROP count also drops to 88 units. Speaking of 11, the memory clock on the G5X implementation on GTX 1080 Ti will now run at 11 Gbps, a boost available to NVIDIA thanks to a chip revision from Micron and improvements to equalization and reverse signal distortion.
The TDP of the new part is 250 watts
, falling between the Titan X and the GTX 1080. That’s an interesting move considering that the GP102 was running at 250 watts with identical to the Titan product. The cooler has been improved compared to the GTX 1080, offering quieter fan speeds and lower temperatures when operating at the same power envelope.
Performance estimates from NVIDIA put the GTX 1080 Ti about 35% faster than the GTX 1080, the largest “kicker performance increase” that we have seen from a flagship Ti launch.
Pricing is going to be set at $699 so don't expect to find this in any budget builds. But for the top performing GeForce card on the market, it's what we expect. It should be on virtual shelves starting next week.
(Side note, with the GTX 1080 getting a $100 price drop tonight, I think we'll find this new lineup very compelling to enthusiasts.)
NVIDIA did finally detail its tiled caching rendering technique. We'll be diving more into that in a separate article with a little more time for research.
One more thing…
In another interesting move, NVIDIA is going to be offering “overclocked” versions of the GTX 1080 and GTX 1060 with +1 Gbps memory speeds. Partners will be offering them with some undisclosed price premium.
I don’t know how much performance this will give us but it’s clear that NVIDIA is preparing its lineup for the upcoming AMD Vega release.
We’ll have more news from NVIDIA and GDC as it comes!
Subject: Graphics Cards | February 28, 2017 - 10:55 PM | Tim Verry
Tagged: pascal, nvidia, GTX 1080, GDC
Update Feb 28 @ 10:03pm It's official, NVIDIA launches $699 GTX 1080 Ti.
NVIDIA is hosting a "Gaming Celebration" live event during GDC 2017 to talk PC gaming and possibly launch new hardware (if rumors are true!). During the event, NVIDIA CEO Jen-Hsun Huang made a major announcement regarding its top-end GTX 1080 graphics card with a price drop to $499 effective immediately.
The NVIDIA GTX 1080 is a pascal based graphics card with 2560 CUDA cores paired with 8GB of GDDR5X memory. Graphics cards based on this GP104 GPU are currently selling for around $580 to $700 (most are around $650+/-) with the "Founders Edition" having an MSRP of $699. The $499 price teased at the live stream represents a significant price drop compared to what the graphics cards are going for now. NVIDIA did not specify if the new $499 MSRP was the new Founders Edition price or an average price that includes partner cards as well but even if it only happened on the reference cards, the partners would have to adjust their prices downwards accordingly to compete.
I suspect that NVIDIA is making such a bold move to make room in their lineup for a new product (the long-rumored 1080 Ti perhaps?) as well as a pre-emptive strike against AMD and their Radeon RX Vega products. This move may also be good news for GTX 1070 pricing as they may also see price drops to make room for cheaper GTX 1080 partner cards that come in below the $499 price point.
If you have been considering buying a new graphics card, NVIDIA has sweetened the pot a bit especially if you had already been eyeing a GTX 1080. (Note that while the price drop is said to be effective immediately, at the time of writing Amazon was still showing "normal"/typical prices for the cards. Enthusiasts might have to wait a few hours or days for the retailers to catch up and update their sites.)
This makes me a bit more excited to see what AMD will have to offer with Vega as well as the likelihood of a GTX 1080 Ti launch happening sooner rather than later!
Subject: General Tech, Graphics Cards | February 27, 2017 - 03:39 PM | Jeremy Hellstrom
Tagged: MWC, GDC, VRMark, Servermark, OptoFidelity, cyan room, benchmark
Futuremark are showing off new benchmarks at GDC and MWC, the two conferences which are both happening this week. We will have quite a bit of coverage this week as we try to keep up with simultaneous news releases and presentations.
First up is a new benchmark in their recently released DX12 VRMark suite, the new Cyan Room which sits between the existing two in the suite. The Orange Room is to test if your system is capable of providing you with an acceptable VR experience or if your system falls somewhat short of the minimum requirements while the Blue Room is to show off what a system that exceeds the recommended specs can manage. The Cyan room will be for those who know that their system can handle most VR, and need to test their systems settings. If you don't have the test suite Humble Bundle has a great deal on this suite and several other tools, if you act quickly.
Next up is a new suite to test Google Daydream, Google Cardboard, and Samsung Gear VR performance and ability. There is more than just performance to test when you are using your phone to view VR content, such as avoiding setting your eyeholes on fire. The tests will help you determine just how long your device can run VR content before overheating becomes an issue and interferes with performance, as well as helping you determine your battery life.
VR Latency testing is the next in the list of announcements and is very important when it comes to VR as high or unstable latency is the reason some users need to add a bucket to their list of VR essentials. Futuremark have partnered with OptoFidelity to produce VR Multimeter HMD hardware based testing. This allows you, and hopefully soon PCPer as well, to test motion-to-photon latency, display persistence, and frame jitter as well as audio to video synchronization and motion-to-audio-latency all of which could lead to a bad time.
Last up is the brand new Servermark to test the performance you can expect out of virtual servers, media servers and other common tasks. The VDI test lets you determine if a virtual machine has been provisioned at a level commensurate to the assigned task, so you can adjust it as required. The Media Transcode portion lets you determine the maximum number of concurrent streams as well as the maximum quality of those streams which your server can handle, very nice for those hosting media for an audience.
Expect to hear more as we see the new benchmarks in action.