Subject: General Tech, Graphics Cards | February 7, 2014 - 12:54 AM | Scott Michaud
Tagged: sli, crossfire
I will not even call this a thinly-veiled rant. Linus admits it. To make a point, he assembled a $5000 PC running a pair of NVIDIA GeForce 780 Ti GPUs and another pair of AMD Radeon R9 290X graphics cards. While Bitcoin mining would likely utilize all four video cards well enough, games will not. Of course, he did not even mention the former application (thankfully).
Honestly, he's right. One of the reasons why I am excited about OpenCL (and its WebCL companion) is that it simply does not care about devices. Your host code manages the application but, when the jobs get dirty, it enlists help from an available accelerator by telling it to perform a kernel (think of it like function) and share the resulting chunk of memory.
This can be an AMD GPU. This can be an NVIDIA GPU. This can be an x86 CPU. This can be an FPGA. If the host has multiple, independent tasks, it can be several of the above (and in any combination). OpenCL really does not care.
Obviously, to be fair, AMD is very receptive to open platforms. NVIDIA is less-so, and they are honest about that, but they conform to standards when it benefits their users more than their proprietary ones. I know that point can be taken multiple ways, and several will be hotly debated, but I really cannot find the words to properly narrow it.
Despite the fragmentation in features, there is one thing to be proud of as a PC gamer. You may have different experiences depending on the components you purchase.
But, at least you will always have an experience.
Subject: General Tech, Graphics Cards | February 6, 2014 - 05:54 PM | Scott Michaud
Tagged: amd, radeon, R7 250X
The AMD Radeon R7 250X has been mentioned on a few different websites over the last day, one of which was tweeted by AMD Radeon Italia. The SKU, which bridges the gap between the R7 250 and the R7 260, is expected to have a graphics processor with 640 Stream Processors, 40 TMUs, and 16 ROPs. It should be a fairly silent launch, with 1GB and 2GB versions appearing soon for an expected price of around 90 Euros, including VAT.
Image Credit: Videocardz.com
The GPU is expected to be based on the 28nm Oland chip design.
While it may seem like a short, twenty Euro jump from the R7 250 to the R7 260, the single-precision FLOP performance actually doubles from around 800 GFLOPs to around 1550 GFLOPs. If that metric is indicative of overall performance, there is quite a large gap to place a product within.
We still do not know official availability, yet.
Subject: General Tech, Graphics Cards | February 6, 2014 - 10:44 AM | Jeremy Hellstrom
We have more news and it is good for Galaxy fans. The newest update states that they will be sticking around!
Good news GPU fans, the rumours that Galaxy's GPU team is leaving the North American market might be somewhat exaggerated, at least according to their PR Team.
This post appeared on Facebook and was quickly taken off again, perhaps for rewording or perhaps it is a perfect example of the lack of communication that [H]ard|OCP cites in their story. Stay tuned as we update you as soon as we hear more.
Party like it's 2008!
[H]ard|OCP have been following Galaxy's business model closely for the past year as they have been seeing hints that the reseller just didn't get the North American market. Their concern grew as they tried and failed to contact Galaxy at the end of 2013, emails went unanswered and advertising campaigns seemed to have all but disappeared. Even with this reassurance that Galaxy is not planning to leave the North American market a lot of what [H] says rings true, with the stock and delivery issues Galaxy seemed to have over the past year there is something going on behind the scenes. Still it is not worth abandoning them completely and turning this into a self fulfilling prophecy, they have been in this market for a long time and may just be getting ready to move forward in a new way. On the other hand you might be buying a product which will not have warranty support in the future.
"The North American GPU market has been one that is at many times a swirling mass of product. For the last few years though, we have seen the waters calm in that regard as video card board partners have somewhat solidified and we have seen solid players emerge and keep the stage. Except now we seen one exit stage left."
Here is some more Tech News from around the web:
- Microsoft's new CEO: The technology isn't his problem @ The Register
- Oculus Releases Open Source Hardware @ Hack a Day
- HP retains the top spot in a declining PC market @ The Inquirer
- Is Intel Selling Bay Trail Chips Below Cost? @ Slashdot
- Lenovo hires ex-CIA bod to push through Moto deal @ The Register
Subject: General Tech, Graphics Cards | February 5, 2014 - 11:43 AM | Jeremy Hellstrom
Tagged: gaming, Mantle, amd, battlefield 4
Now that the new Mantle enabled driver has been released several sites have had a chance to try out the new API to see what effect it has on Battlefield 4. [H]ard|OCP took a stock XFX R9 290X paired with an i7-3770K and tested both single and multiplayer BF4 performance and the pattern they saw lead them to believe Mantle is more effective at relieving CPU bottlenecks than ones caused by the GPU. The performance increases they saw were greater at lower resolutions than at high resolutions. At The Tech Report another XFX R9 290X was paired with an A10-7850K and an i7-4770K and compared the systems performance in D3D as well as Mantle. To make the tests even more interesting they also tested D3D with a 780Ti, which you should fully examine before deciding which performs the best. Their findings were in line with [H]ard|OCP's and they made the observation that Mantle is going to offer the greatest benefits to lower powered systems, with not a lot to be gained by high end systems with the current version of Mantle. Legit Reviews performed similar tests but also brought the Star Swarm demo into the mix, using an R7 260X for their GPU. You can catch all of our coverage by clicking on the Mantle tag.
"Does AMD's Mantle graphics API deliver on its promise of smoother gaming with lower-spec CPUs? We take an early look at its performance in Battlefield 4."
Here is some more Tech News from around the web:
- Humble Sid Meier Bundle announced: So much Civilisation! @ HEXUS
- HARD ONES: Three new PC games that are BLOODY DIFFICULT @ The Register
- Developers Reporting No Payments From Strategy First @ Rock, Paper, SHOTGUN
What Mantle signifies about GPU architectures
Mantle is a very interesting concept. From the various keynote speeches, it sounds like the API is being designed to address the current state (and trajectory) of graphics processors. GPUs are generalized and highly parallel computation devices which are assisted by a little bit of specialized silicon, when appropriate. The vendors have even settled on standards, such as IEEE-754 floating point decimal numbers, which means that the driver has much less reason to shield developers from the underlying architectures.
Still, Mantle is currently a private technology for an unknown number of developers. Without a public SDK, or anything beyond the half-dozen keynotes, we can only speculate on its specific attributes. I, for one, have technical questions and hunches which linger unanswered or unconfirmed, probably until the API is suitable for public development.
Or, until we just... ask AMD.
Our response came from Guennadi Riguer, the chief architect for Mantle. In it, he discusses the API's usage as a computation language, the future of the rendering pipeline, and whether there will be a day where Crossfire-like benefits can occur by leaving an older Mantle-capable GPU in your system when purchasing a new, also Mantle-supporting one.
Q: Mantle's shading language is said to be compatible with HLSL. How will optimizations made for DirectX, such as tweaks during shader compilation, carry over to Mantle? How much tuning will (and will not) be shared between the two APIs?
[Guennadi] The current Mantle solution relies on the same shader generation path games the DirectX uses and includes an open-source component for translating DirectX shaders to Mantle accepted intermediate language (IL). This enables developers to quickly develop Mantle code path without any changes to the shaders. This was one of the strongest requests we got from our ISV partners when we were developing Mantle.
Follow-Up: What does this mean, specifically, in terms of driver optimizations? Would AMD, or anyone else who supports Mantle, be able to re-use the effort they spent on tuning their shader compilers (and so forth) for DirectX?
[Guennadi] With the current shader compilation strategy in Mantle, the developers can directly leverage DirectX shader optimization efforts in Mantle. They would use the same front-end HLSL compiler for DX and Mantle, and inside of the DX and Mantle drivers we share the shader compiler that generates the shader code our hardware understands.
Subject: General Tech, Graphics Cards, Processors | February 4, 2014 - 11:08 PM | Scott Michaud
Tagged: photoshop, opencl, Adobe
Adobe has recently enhanced Photoshop CC to accelerate certain filters via OpenCL. AMD contacted NitroWare with this information and claims of 11-fold performance increases with "Smart Sharpen" on Kaveri, specifically. The computer hardware site decided to test these claims on a Radeon HD 7850 using the test metrics that AMD provided them.
Sure enough, he noticed a 16-fold gain in performance. Without OpenCL, the filter's loading bar was on screen for over ten seconds; with it enabled, there was no bar.
Dominic from NitroWare is careful to note that an HD 7850 is significantly higher performance than an APU (barring some weird scenario involving memory transfers or something). This might mark the beginning of Adobe's road to sensible heterogeneous computing outside of video transcoding. Of course, this will also be exciting for AMD. While they cannot keep up with Intel, thread per thread, they are still a heavyweight in terms of total performance. With Photoshop, people might actually notice it.
Subject: General Tech, Graphics Cards | February 1, 2014 - 08:29 PM | Scott Michaud
Tagged: Mantle, BF4, amd
AMD has released the Catalyst 14.1 Beta driver (even for Linux) but you should, first, read Ryan's review. This is a little less than what he expects in a Beta from AMD. We are talking about crashes to desktop and freezes while loading a map on a single-GPU configuration - and Crossfire is a complete wash in his experience (although AMD acknowledges the latter in their release notes). According to AMD, there is even the possibility that the Mantle version of Battlefield 4 will render with your APU and ignore your dedicated graphics.
If you are determined to try Catalyst 14.1, however, it does make a first step into the promise of Mantle. Some situations show slightly lower performance than DirectX 11, albeit with a higher minimum framerate, while other results impress with double-digit percentage gains.
Multiplayer in BF4, where the CPU is more heavily utilized, seems to benefit the most (thankfully).
If you understand the risk (in terms of annoyance and frustration), and still want to give it a try, pick up the driver from AMD's support website. If not? Give it a little more time for AMD to whack-a-bug. At some point, there should be truly free performance waiting for you.
A quick look at performance results
Late last week, EA and Dice released the long awaited patch for Battlefield 4 that enables support for the Mantle renderer. This new API technology was introduced by AMD back in September. Unfortunately, AMD wasn't quite ready for its release with their Catalyst 14.1 beta driver. I wrote a short article that previewed the new driver's features, its expected performance with the Mantle version of BF4, and commentary about the current state of Mantle. You should definite read that as a primer before continuing if you haven't yet.
Today, after really just a few short hours with a useable driver, I have only limited results. Still, I know that you, our readers, clamor for ANY information on the topic. I thought I would share what we have thus far.
As I mentioned in the previous story, the Mantle version of Battlefield 4 has the biggest potential to show advantages in times where the game is more CPU limited. AMD calls this the "low hanging fruit" for this early release of Mantle and claim that further optimizations will come, especially for GPU-bound scenarios. Because of that dependency on CPU limitations, that puts some non-standard requirements on our ability to showcase Mantle's performance capabilities.
For example, the level of the game and even the section of that level, in the BF4 single player campaign, can show drastic swings in Mantle's capabilities. Multiplayer matches will also show more consistent CPU utilization (and thus could be improved by Mantle) though testing those levels in a repeatable, semi-scientific method is much more difficult. And, as you'll see in our early results, I even found a couple instances in which the Mantle API version of BF4 ran a smidge slower than the DX11 instance.
For our testing, we compiled two systems that differed in CPU performance in order to simulate the range of processors installed within consumers' PCs. Our standard GPU test bed includes a Core i7-3960X Sandy Bridge-E processor specifically to remove the CPU as a bottleneck and that has been included here today. We added in a system based on the AMD A10-7850K Kaveri APU which presents a more processor-limited (especially per-thread) system, overall, and should help showcase Mantle benefits more easily.
Subject: Graphics Cards, Processors | January 31, 2014 - 01:36 PM | Ryan Shrout
Tagged: 7850k, A10-7850K, amd, APU, gt 630, Intel, nvidia, video
As a follow up to our first video posted earlier in the week that looked at the A10-7850K and the GT 630 from NVIDIA in five standard games, this time we compare the A10-7850K APU against the same combination of the Intel and NVIDIA hardware in five of 2013's top free to play games.
UPDATE: I've had some questions about WHICH of the GT 630 SKUs were used in this testing. Our GT 630 was this EVGA model that is based on 96 CUDA cores and a 128-bit DDR3 memory interface. You can see a comparison of the three current GT 630 options on NVIDIA's website here.
If you are looking for more information on AMD's Kaveri APUs you should check out my review of the A8-7600 part as well our testing of Dual Graphics with the A8-7600 and a Radeon R7 250 card.
Subject: Graphics Cards | January 30, 2014 - 12:15 PM | Jeremy Hellstrom
Tagged: xfx, double d, R9 290X
The only thing more fun that an XFX Double Dissipation R9 290X is two of them in Crossfire, which is exactly what [H]ard|OCP just tested. These cards sport the familiar custom cooler though they are not overclocked nor is [H] testing overclocking in this review though they will revisit this card in the future to do exactly that. This review is about the Crossfire performance of these cards straight out of the box and it is rather impressive. When [H] tested 4K performance they could feel the frame pacing improvements the new driver gives as well as seeing these cards outperform the SLI'd GTX 780 Ti cards in every test; though not always by a huge margin. The current selling price of these cards is about $100 above the MSRP but still come in cheaper than the current NVIDIA card; these particular cards really show off what Hawaii can be capable of.
"Take two custom XFX R9 290X Double Dissipation Edition video cards, enable CrossFire, and let your jaw hit the floor. We will test this combination against the competition in a triple-display Eyefinity setup as well as 4K Ultra HD display gaming. We will find out if custom cards hold any advantage over the reference designed R9 290X."
Here are some more Graphics Card articles from around the web:
- Sapphire Radeon R7 260X @ Phoronix
- Gigabyte R9 290 WindForce OC @ Kitguru
- XFX Radeon R9 280X Black Edition @ Benchmark Reviews
- XFX Radeon R9 290X Double Dissipation Review @ Hardware Canucks
- Gigabyte R9 290X WindForce OC 4GB @ eTeknix
- Sapphire Dual-X R9 270 Graphics Card Review @ Modders-Inc
- EK Waterblocks R280X Matrix Edition Full Cover Block Review @ Madshrimps
- 24-Way AMD Radeon vs. NVIDIA GeForce Linux Graphics Card Comparison @ Phoronix
- 25-Way Open-Source Linux Graphics Card Comparison @ Phoronix
- MSI GTX 760 Mini-ITX Gaming 2 GB @ techPowerUp
Get notified when we go live!