All | Editorial | General Tech | Graphics Cards | Networking | Motherboards | Cases and Cooling | Processors | Chipsets | Memory | Displays | Systems | Storage | Mobile | Shows and Expos
Introduction and Technical Specifications
Courtesy of Cooler Master
Cooler Master is known in the enthusiast community for their innovative designs with product offerings ranging from cases to desktop and laptop cooling implements. Cooler Master also offers their own line of all-in-one (AIO) CPU liquid cooling solutions for better system performance without the noise of a typical air cooler. With their Nepton 240M cooler, they enhanced the existing design of their previous AIO products, optimizing its performance with an enhanced pump and radiator design. We measured the unit's performance against that of other high-performance liquid and air coolers to best illustrate its abilities. The Nepton 240M's premium performance comes with a premium price, at a $139.99 MSRP.
Courtesy of Cooler Master
Courtesy of Cooler Master
The Nepton 240M AIO liquid cooler features a 240mm aluminum-finned radiator tied to a base unit consisting of a 120 liter per minute pump and a micro-finned copper base plate. Unlike the Glacer model, the Nepton 240M does not feature the ability to drain and refill the unit. Cooler Master designed the Nepton 240M with a 27mm deep, 2x120mm copper radiator with brass internal channels, bundled with two of its 120mm Silencio model fans. The Silencio fans are optimized for low noise and high pressure, perfect for use with a liquid cooling radiator. The radiator and unit base are connected by ribbed FEP (Fluorinated Ethylene Propylene) tubing, allowing for high flexibility without the worry of tube kinking.
Quiet, Efficient Gaming
The last few weeks have been dominated by talk about the memory controller of the Maxwell based GTX 970. There are some very strong opinions about that particular issue, and certainly NVIDIA was remiss on actually informing consumers about how it handles the memory functionality of that particular product. While that debate rages, we have somewhat lost track of other products in the Maxwell range. The GTX 960 was released during this particular firestorm and, while it also shared the outstanding power/performance qualities of the Maxwell architecture, it is considered a little overpriced when compared to other cards in its price class in terms of performance.
It is easy to forget that the original Maxwell based product to hit shelves was the GTX 750 series of cards. They were released a year ago to some very interesting reviews. The board is one of the first mainstream cards in recent memory to have a power draw that is under 75 watts, but can still play games with good quality settings at 1080P resolutions. Ryan covered this very well and it turned out to be a perfect gaming card for many pre-built systems that do not have extra power connectors (or a power supply that can support 125+ watt graphics cards). These are relatively inexpensive cards and very easy to install, producing a big jump in performance as compared to the integrated graphics components of modern CPUs and APUs.
The GTX 750 and GTX 750 Ti have proven to be popular cards due to their overall price, performance, and extremely low power consumption. They also tend to produce a relatively low amount of heat, due to solid cooling combined with that low power consumption. The Maxwell architecture has also introduced some new features, but the major changes are to the overall design of the architecture as compared to Kepler. Instead of 192 cores per SMK, there are now 128 cores per SMM. NVIDIA has done a lot of work to improve performance per core as well as lower power in a fairly dramatic way. An interesting side effect is that the CPU hit with Maxwell is a couple of percentage points higher than Kepler. NVIDIA does lean a bit more on the CPU to improve overall GPU power, but most of this performance hit is covered up by some really good realtime compiler work in the driver.
Asus has taken the GTX 750 Ti and applied their STRIX design and branding to it. While there are certainly faster GPUs on the market, there are none that exhibit the power characteristics of the GTX 750 Ti. The combination of this GPU and the STRIX design should result in an extremely efficient, cool, and silent card.
Project Lead: Joris-Jan van ‘t Land
Thanks to Ian Comings, guest writer from the PC Perspective Forums who conducted the interview of Bohemia Interactive's Joris-Jan van ‘t Land. If you are interested in learning more about ArmA 3 and hanging out with some PC gamers to play it, check out the PC Perspective Gaming Forum!
I recently got the chance to send some questions to Bohemia Interactive, a computer game development company based out of Prague, Czech Republic, and a member of IDEA Games. Bohemia Interactive was founded in 1999 by CEO Marek Španěl, and it is best known for PC gaming gems like Operation Flashpoint: Cold War Crisis, The ArmA series, Take On Helicopters, and DayZ. The questions are answered by ArmA 3's Project Lead: Joris-Jan van ‘t Land.
PC Perspective: How long have you been at Bohemia Interactive?
VAN ‘T LAND: All in all, about 14 years now.
PC Perspective: What inspired you to become a Project Lead at Bohemia Interactive?
VAN ‘T LAND: During high school, it was pretty clear to me that I wanted to work in game development, and just before graduation, a friend and I saw a first preview for Operation Flashpoint: Cold War Crisis in a magazine. It immediately looked amazing to us; we were drawn to the freedom and diversity it promised and the military theme. After helping run a fan website (Operation Flashpoint Network) for a while, I started to assist with part-time external design work on the game (scripting and scenario editing). From that point, I basically grew naturally into this role at Bohemia Interactive.
PC Perspective: What part of working at Bohemia Interactive do you find most satisfying? What do you find most challenging?
VAN ‘T LAND: The amount of freedom and autonomy is very satisfying. If you can demonstrate skills in some area, you're welcome to come up with random ideas and roll with them. Some of those ideas can result in official releases, such as Arma 3 Zeus. Another rewarding aspect is the near real-time connection to those people who are playing the game. Our daily Dev-Branch release means the work I do on Monday is live on Tuesday. Our own ambitions, on the other hand, can sometimes result in some challenges. We want to do a lot and incorporate every aspect of combat in Arma, but we're still a relatively small team. This can mean we bite off more than we can deliver at an acceptable level of quality.
PC Perspective: What are some of the problems that have plagued your team, and how have they been overcome?
VAN ‘T LAND: One key problem for us was that we had no real experience with developing a game in more than one physical location. For Arma 3, our team was split over two main offices, which caused quite a few headaches in terms of communication and data synchronization. We've since had more key team members travel between the offices more frequently and improved our various virtual communication methods. A lot of work has been done to try to ensure that both offices have the latest version of the game at any given time. That is not always easy when your bandwidth is limited and games are getting bigger and bigger.
AMD Details Carrizo Further
Some months back AMD introduced us to their “Carrizo” product. Details were slim, but we learned that this would be another 28 nm part that has improved power efficiency over its predecessor. It would be based on the new “Excavator” core that will be the final implementation of the Bulldozer architecture. The graphics will be based on the latest iteration of the GCN architecture as well. Carrizo would be a true SOC in that it integrates the southbridge controller. The final piece of information that we received was that it would be interchangeable with the Carrizo-L SOC, which is a extremely low power APU based on the Puma+ cores.
A few months later we were invited by AMD to their CES meeting rooms to see early Carrizo samples in action. These products were running a variety of applications very smoothly, but we were not informed of speeds and actual power draw. All that we knew is that Carrizo was working and able to run pretty significant workloads like high quality 4K video playback. Details were yet again very scarce other than the expected timeline of release, the TDP ratings of these future parts, and how it was going to be a significant jump in energy efficiency over the previous Kaveri based APUs.
AMD is presenting more information on Carrizo at the ISSCC 2015 conference. This information dives a little deeper into how AMD has made the APU smaller, more power efficient, and faster overall than the previous 15 watt to 35 watt APUs based on Kaveri. AMD claims that they have a product that will increase power efficiency in a way not ever seen before for the company. This is particularly important considering that Carrizo is still a 28 nm product.
Intel Pushes Broadwell to the Next Unit of Computing
Intel continues to invest a significant amount of money into this small form factor product dubbed the Next Unit of Computing, or NUC. When it was initially released in December of 2012, the NUC was built as an evolutionary step of the desktop PC, part of a move for Intel to find new and unique form factors that its processors can exist in. With a 4" x 4" motherboard design the NUC is certainly a differentiating design and several of Intel's partners have adopted it for products of their: Gigabyte's BRIX line being the most relevant.
But Intel's development team continues to push the NUC platform forward and today we are evaluating the most recent iteration. The Intel NUC5i5RYK is based on the latest 14nm Broadwell processor and offers improved CPU performance, a higher speed GPU and lower power consumption. All of this is packed into a smaller package than any previous NUC on the market and the result is both impressive and totally expected.
A Walk Around the NUC
To most poeple the latest Intel NUC will look very similar to the previous models based on Ivy Bridge and Haswell. You'd be right of course - the fundamental design is unchanged. But Intel continues to push forward in small ways, nipping and tucking away. But the NUC is still just a box. An incredibly small one with a lot of hardware crammed into it, but a box none the less.
While I can appreciate the details including the black and silver colors and rounded edges, I think that Intel needs to find a way to add some more excitement into the NUC product line going forward. Admittedly, it is hard to inovate in that directions with a focus on size and compression.
Just over a week or so ago Allyn spent some time with the MSI X99A Gaming 9 ACK motherboard, a fact that might seem a little odd to our frequent readers. Why would our storage editor be focusing on a motherboard? USB 3.1 of course! When we visited MSI at CES in January they were the first company to show working USB 3.1 hardware and performance numbers that we were able duplicate in our testing when MSI sent us similar hardware.
But ASUS is in this game as well, preparing its product lines with USB 3.1 support courtesy of the same ASMedia controller we looked at before. ASUS has a new revision of several motherboards planned with integrated on-board USB 3.1 but is also going to be releasing an add-in card with USB 3.1 support for existing systems.
Today we are going to test that add-in card to measure ASUS' implementation of USB 3.1 and see how it stacks up to what MSI had to offer and what improvements and changes you can expect from USB 3.0.
USB 3.1 Technology Background
Despite the simple point denomination change in USB 3.1, also known as SuperSpeed+, the technological and speed differences in the newest revision of USB are substantial. Allyn did a good job of summarizing the changes that include a 10 Gbps link interface and a dramatic drop in encoding overhead that enables peak theoretical performance improvements of 2.44x compared to USB 3.0.
USB 3.1 is rated at 10 Gbps, twice that of USB 3.0. The little-reported-on nugget of info from the USB 3.1 specification relates to how they classify the raw vs. expected speeds. Taking USB 3.0 as an example, Superspeed can handle a raw 5Gbps data rate, but after subtracting out the overhead (packet framing, flow control, etc), you are left with ~450MB/s of real throughput. Superspeed+ upgrades the bit encoding type from 8b/10b (80% efficient) to 128b/132b (97% efficient) *in addition to* the doubling of raw data rate. This means that even after accounting for overhead, Superspeed+’s best case throughput should work out to ~1.1GB/s. That’s not a 2x speed improvement – it is actually 2.44x of USB 3.0 speed. Superspeed+ alright!
The ASUS STRIX TACTIC PRO is a premium mechanical gaming keyboard featuring Cherry MX Brown switches and some serious style.
Keyboards are a very personal thing, and as this is one of the three primary interfaces with the system itself (along with the mouse and display), feel will help decide the experience. Without a doubt mechanical keyboard have become very popular with enthusiasts, but as more manufacturers have started offering them - and the market has begun to saturate - it becomes much more difficult to pick a starting point if you're new to the game. To further complicate a buying decision there are different types of key switches used in these keyboards, and each variety has its own properties and unique feel.
And on the subject of key switches, this particular keyboard built with the brown variety of the Cherry MX switches, and ASUS offers the option of Cherry MX Black, Blue, and Red switches with the STRIX TACTIC PRO as well. Our own Scott Michaud covered the topic of key switches in great detail last year, and that article is a great starting point that helps explain the different types of switches available, and how they differ.
The Cherry MX Brown switch in action
I'll go into the feel of the keyboard on the next page, but quickly I'll say that MX Brown switches have a good feel without being too "clicky", but they are certainly more stiff feeling than a typical membrane keyboard. While it's impossible to really describe how the keyboard will feel to a particular user, we can certainly cover the features and performance of this keyboard to help with a purchasing decision in this crowded market. At $150 the STRIX TACTIC PRO carries a premium price, but as you'll see this is also a premium product.
Flagship. Premium. Best in class. These are the terms that Dell and Intel muttered to me during a conference call to discuss the new Dell Venue 8 7000 tablet. It’s a bullish claim and one that would likely have been received with a sideways eye roll or a shrug had I not been able to get a short amount of hands on time with the device at CES in January. The idea that Dell would develop an Android tablet that bests what more established brands like Nexus and Samsung have created, AND that that same tablet would be powered by an Intel processor rather than a Qualcomm, NVIDIA or Samsung chip would have seemed laughable last year. But after a solid three weeks with the Venue 8 7000 I am prepared to make the statement: this is my favorite tablet. Not my favorite Intel tablet, not my favorite Android tablet: just plain favorite.
The Venue 8 7000 combines style, design, technology and visuals that are simply unmatched by anything else in the Android word and rivals anything that Apple has created to date. There are a couple of warts that center around the camera and gaming performance that won’t drop your jaw, but for the majority of use cases the user experience is as exceptional as the looks.
Maybe best of all, this tablet starts at just $399 and is available today.
Dell Venue 8 7000 Specifications
Let’s begin the review by looking at the raw specifications of the Dell Venue 8 7000. Even though hardware specifications don’t tell a complete story of any device, especially a tablet that is based so much on experience, it is important to get a good baseline expectation.
|Dell Venue 8 7000 (Model 7840)|
|Processor||Intel Atom Z3580 Quad-Core 2.33 GHz|
|Screen||2560x1600 OLED 8.4-in (359 ppi)|
MicroSD Slot (up to 512GB)
|Camera||8MP Rear + Dual 720p Depth
|Wireless||Intel 7260 802.11ac 1x1 Dual Band
|Connection||USB 2.0 (power and data)
|Dimensions||215.8mm x 124.4mm x 6mm
8.5" x 4.88" x 0.24"
The center of the Venue 8 7000 is the Intel Atom Z3580 quad-core processor with a peak clock rate of 2.3 GHz and a base clock rate of 500 MHz. The Z3580 is a 22nm processor based on the Moorefield platform and Silvermont architecture. I first got information about the Silvermont architecture back in May of 2013 so it seems a bit dated in some regards, but the performance and power efficiency is still there to compete with the rival options from ARM.. The Venue 8 7000 includes an LPDDR3-1600 controller and there is 2GB of memory; a decent amount but we are seeing quite a few smartphones with more system memory like the OnePlus One.
Introduction and Features
Our first Corsair power supply up for review in 2015 is the CS Series Modular 850W PSU; the CS850M. Corsair's CS Series Modular PSUs are designed for basic desktop use and light to moderate gaming where low energy use, low noise, simple installation, and good value are important. The Modular CS Series now includes five models; the CS450M, CS550M, CS650M, CS750M, and the new CS850M. All of the power supplies in the CS Series feature modular cables, high efficiency (80 Plus Gold certified) and quiet operation. In addition, Corsair continues to offer a full line of high quality power supplies, memory components, cases, cooling components, SSDs and accessories for the PC market.
Here is what Corsair has to say about their CS Series Modular PSUs: “The CS-M Series is designed for basic and midrange PCs, but offers features and performance traditionally reserved for higher-end models. 80 Plus Gold efficiency and a thermally controlled fan ensure quiet operation and lower energy use, and the modular, detachable cable set makes installations and upgrades faster and better looking.”
“80 Plus Gold efficiency reduces operating cost and excess heat. Since it generates less heat, the fan doesn’t need to work as hard, and you’ll enjoy near silent operation. The flat black modular cables with clearly-marked connectors make installation fast and straightforward, with good-looking results.”
Corsair CS Series Modular PSU Key Features:
• Five Models: 450W up to 850W
• Compliant with the latest ATX12V v2.4 and EPS 2.92 standards
• Backward compatible with ATX12V 2.2, 2.31 and ATX12V 2.01 systems
• 4th Generation Intel® Core™ processor ready (Haswell & Z87 motherboards)
• 80 Plus Gold certified for high efficiency (=90% under real world loads)
• Modular cables (only use the cables you need)
• Low-profile, flat modular cables reduce air friction and maximize airflow
• Active PFC with Universal AC input (100-240VAC)
• Multi GPU ready
• Safety: OVP, UVP, SCP, OPP, and OTP
• Approvals: FCC, ICES, UL, CUL, TUV, CCC, CE, RCM, CB, EAC, KC, BSMI, ROHS, WEEE
• 3-Year warranty and lifetime access to Corsair’s tech support & customer service
• MSRP: $139.99 USD
Introduction, Specifications and Packaging
Micron's Crucial brand has been cranking out some great low cost SSDs for the past several years now. While their early drives pushed into the SATA 6Gb/sec interface before most of the competition, their performance was inconsistent and lagged behind some of the other more nimble solutions available at that time. This pattern was broken around the time of the M550 and MX100 launches. Those two drives were heavily competitive in performance and even moreso in pricing. Actually the pricing is probably the bigger story - when they launched, one of our readers caught a 512GB MX100 on sale for $125 ($0.24/GB)! We are coming up on a year since the MX100, and at CES 2015 Micron launched a pair of SSD models - the BX100 and MX200. Today we are going to look at the BX100 series:
Crucial aims to make the BX100 as their lowest cost/GB SSD ever - even cheaper than the MX100. Since Micron makes the flash, the best way to drive costs down is to use a lower cost controller. The Silicon Motion SM2246EN is cheaper to procure than the equivalent Marvell part, yet still performs rather well.
The Silicon Motion SM2246EN SSD controller
This is a great controller, as we have seen in our prior review of the ADATA SP610, Corsair Neutron LX, and Angelbird SSD WRK. From the specs, we can see that Micron has somehow infused their variant with increased write speeds even though it appears to use the same flash as those competing models listed above. We'll see how this plays out as the review progresses.
Introduction: This Is Not a NAS
The new WSS NAS series from Thecus contains some very interesting devices, and particularly so at the entry-level price with the unit we’re looking at today. WSS is the abbreviation for Windows Storage Server (in this case it’s 2012 R2), and this provides a huge increase in functionality compared to a standard NAS, as you might imagine.
Need a server? Just add a keyboard, mouse, and monitor
It’s really quite remarkable what Thecus is doing in partnership with Microsoft here in terms of value, as this entry 2-bay unit costs just $350. While this may seem high for a dual-bay NAS, we really aren’t talking about a NAS at all with this - which will be readily apparent to the user upon first powering it up. We are talking about a full-scale server here, replete with Windows Server 2012 R2 Essentials goodness. Of course a savvy user could easily deploy a small server in a home or office, and there are many advantages to managed solutions beyond the simple NAS appliances. But the advantage of a NAS is just that: it is significantly less complex and accessible for a consumer. The W2000 presents a very interesting option due to one particular aspect of its own accessibility: price. At $350 you are getting a very compact server with internal hardware much more akin to a standard desktop than you might imagine, and it ships installed with Microsoft's Windows Storage Server 2012 R2 Essentials.
What is “Storage” Server Essentials?
Ok, so I was a little confused as to the specific difference with the Storage version of the Server OS, unless it was simply a licensing distinction. My research first brought me to this quote from Microsoft:
“Windows Storage Server 2012 R2 Essentials is based on Windows Server 2012 R2. In fact, when it comes to functionality, you get key some features that aren’t included in these first two editions.”
After looking through the available documentation it appears as though Storage Server Essentials is, essentially, just Server Essentials with the distinction of being licensed differently. Microsoft TechNet defines it further:
“A computer that runs Windows Storage Server is referred to as a storage appliance. Windows Storage Server is based on the Windows Server operating system, and it is specifically optimized for use with network-attached storage devices. Windows Storage Server offers you a platform to build storage appliances that are customized for your hardware.”
Introduction and Background
We first got a peek of USB 3.1 at CES 2015. MSI had a cool demo showing some throughput figures including read and write speeds as high as 690 MB/s, well over the ~450 MB/s we see on USB 3.0 options shipping today.
We were of course eager to play around with this for ourselves, and MSI was happy to oblige, sending along a box of goodies:
Stuff we will be testing today (Samsung T1 was not part of the MSI demo).
For those unaware, USB 3.1 (also known as Superspeed+), while only a 0.1 increment in numbering, incorporates a doubling of raw throughput and some dramatic improvements to the software overhead of the interface.
Don't be confused between the USB 3.1 standard and the new USB Type-C connector - they are unrelated and independent of each other.
Yes, you’re all going to have to buy *more* cables in the future.
Type-C connectors will enable more simple cable design and thinner connections going forward but USB 3.1 will exist in both Type-A/B and Type-C going forward. Our benchmarking today will utilize Type-A.
New Features and Specifications
It is increasingly obvious that in the high end smartphone and tablet market, much like we saw occur over the last several years in the PC space, consumers are becoming more concerned with features and experiences than just raw specifications. There is still plenty to drool over when looking at and talking about 4K screens in the palm of your hand, octa-core processors and mobile SoC GPUs measuring performance in hundreds of GFLOPS, but at the end of the day the vast majority of consumers want something that does something to “wow” them.
As a result, device manufacturers and SoC vendors are shifting priorities for performance, features and how those are presented both the public and to the media. Take this week’s Qualcomm event in San Diego where a team of VPs, PR personnel and engineers walked me through the new Snapdragon 810 processor. Rather than showing slide after slide of comparative performance numbers to the competition, I was shown room after room of demos. Wi-Fi, LTE, 4K capture and playback, gaming capability, thermals, antennae modifications, etc. The goal is showcase the experience of the entire platform – something that Qualcomm has been providing for longer than just about anyone in this business, while educating consumers on the need for balance too.
As a 15-year veteran of the hardware space my first reaction here couldn’t have been scripted any more precisely: a company that doesn’t show performance numbers has something to hide. But I was given time with a reference platform featuring the Snapdragon 810 processor in a tablet form-factor and the results show impressive increases over the 801 and 805 processors from the previous family. Rumors of the chips heat issues seem overblown, but that part will be hard to prove for sure until we get retail hardware in our hands to confirm.
Today’s story will outline the primary feature changes of the Snapdragon 810 SoC, though there was so much detail presented at the event with such a short window of time for writing that I definitely won’t be able to get to it all. I will follow up the gory specification details with performance results compared to a wide array of other tablets and smartphones to provide some context to where 810 stands in the market.
Introduction and Features
Last month we took a look at SilverStone’s small form-factor power supply the SFX-600, which delivered 600 watts from a compact SFX enclosure. Today we are looking at SilverStone’s new Strider Gold ST1500-GS, which is a 1,500 watt ATX form-factor power supply. The Strider Gold 1500W PSU is fully modular and built for high efficiency operation. What makes the ST1500-GS unique is the relatively short enclosure, which is only 180mm (7.1”) deep!
SilverStone ST1500-GS ATX Power Supply
There are currently five different models available in the Strider Gold S Series, which include the ST55F-G, ST65F-G, ST75-GS, ST85F-GS, and ST1500-GS. All of the Strider Gold S Series PSUs are designed to be fully modular, 80 Plus Gold certified, and small in size. While the typical 1500W power supply enclosure measures 220mm (8.7”) deep, the Strider Gold ST1500-GS is housed in a 180mm chassis.
(Courtesy of SilverStone)
SilverStone Strider Gold S Series ST1500-GS PSU Key Features:
• 1,500 watts DC power output (1600W peak power)
• High efficiency with 80 Plus Gold certification
• 100% Modular cables
• 24/7 Continuous power output with 40°C operating temperature
• Strict ±3% voltage regulation and low AC ripple & noise
• Dedicated single +12V rail (125A)
• Quiet 135mm ball bearing fan
• Eight PCI-E 8/2-pin connectors support multiple high-end graphic adapters
• Conforms to ATX12V and EPS standards
• Universal AC input (100-240V) with Active PFC
• Dimensions: 150mm (W) x 86mm (H) x 180mm (L)
• $319.99 USD
Introduction, Specifications and Packaging
Plextor launched their M6e PCIe SSD in mid-2014. This was the first consumer retail available native PCIe SSD. While previous solutions such as the OCZ RevoDrive bridged SATA SSD controllers to PCIe through a RAID or VCA device, the M6e went with a Marvell controller that could speak directly to the host system over a PCIe 2.0 x2 link. Since M.2 was not widely available at launch time, Plextor also made the M6e available with a half-height PCIe interposer, making for a painless upgrade for those on older non M.2 motherboards (which at that time was the vast majority).
With the M6e out for only a few months time (and in multiple versions), I was surprised to see Plextor launch an additonal version of it at the 2015 CES this past January. Announced alongside the upcoming M7e, the M6e Black Edition is essentially a pimped out version of the original M6e PCIe:
We left CES with a sample of the M6e Black, but had to divert our attention to a few other pressing issues shortly after. With all of that behind us, it's time to get back to cranking out the storage goodness, so let's get to it!
A baker's dozen of GTX 960
Back on the launch day of the GeForce GTX 960, we hosted NVIDIA's Tom Petersen for a live stream. During the event, NVIDIA and its partners provided ten GTX 960 cards for our live viewers to win which we handed out through about an hour and a half. An interesting idea was proposed during the event - what would happen if we tried to overclock all of the product NVIDIA had brought along to see what the distribution of results looked like? After notifying all the winners of their prizes and asking for permission from each, we started the arduous process of testing and overclocking a total of 13 (10 prizes plus our 3 retail units already in the office) different GTX 960 cards.
Hopefully we will be able to provide a solid base of knowledge for buyers of the GTX 960 that we don't normally have the opportunity to offer: what is the range of overclocking you can expect and what is the average or median result. I think you will find the data interesting.
The 13 Contenders
Our collection of thirteen GTX 960 cards includes a handful from ASUS, EVGA and MSI. The ASUS models are all STRIX models, the EVGA cards are of the SSC variety, and the MSI cards include a single Gaming model and three 100ME. (The only difference between the Gaming and 100ME MSI cards is the color of the cooler.)
To be fair to the prize winners, I actually assigned each of them a specific graphics card before opening them up and testing them. I didn't want to be accused of favoritism by giving the best overclockers to the best readers!
SFF PCs get an upgrade
Ultra compact computers, otherwise known as small form factor PCs, are a rapidly increasing market as consumers realize that, for nearly all purposes other than gaming and video editing, Ultrabook-class hardware is "fast enough". I know that some of our readers will debate that fact, and we welcome the discussion, but as CPU architectures continue to improve in both performance and efficiency, you will be able to combine higher performance into smaller spaces. The Gigabyte BRIX platform is the exact result that you expect to see with that combination.
Previously, we have seen several other Gigabyte BRIX devices including our first desktop interaction with Iris Pro graphics, the BRIX Pro. Unfortunately though, that unit was plagued by noise issues - the small fan spun pretty fast to cool a 65 watt processor. For a small computer that would likely sit on top of your desk, that's a significant drawback.
Intel Ivy Bridge NUC, Gigabyte BRIX S Broadwell, Gigabyte BRIX Pro Haswell
This time around, Gigabyte is using the new Broadwell-U architecture in the Core i7-5500U and its significantly lower, 15 watt TDP. That does come with some specification concessions though, including a dual-core CPU instead of a quad-core CPU and a peak Turbo clock rate that is 900 MHz lower. Comparing the Broadwell BRIX S to the more relevant previous generation based on Haswell, we get essentially the same clock speed, a similar TDP, but also an improved core architecture.
Today we are going to look at the new Gigabyte BRIX S featuring the Core i7-5500U and an NFC chip for some interesting interactions. The "S" designates that this model could support a full size 2.5-in hard drive in addition to the mSATA port.
ARM Releases Top Cortex Design to Partners
ARM has an interesting history of releasing products. The company was once in the shadowy background of the CPU world, but with the explosion of mobile devices and its relevance in that market, ARM has had to adjust how it approaches the public with their technologies. For years ARM has announced products and technology, only to see it ship one to two years down the line. It seems that with the increased competition in the marketplace from Apple, Intel, NVIDIA, and Qualcomm ARM is now pushing to license out its new IP in a way that will enable their partners to achieve a faster time to market.
The big news this time is the introduction of the Cortex A72. This is a brand new design that will be based on the ARMv8-A instruction set. This is a 64 bit capable processor that is also backwards compatible with 32 bit applications programmed for ARMv7 based processors. ARM does not go into great detail about the product other than it is significantly faster than the previous Cortex-A15 and Cortex-A57.
The previous Cortex-A15 processors were announced several years back and made their first introduction in late 2013/early 2014. These were still 32 bit processors and while they had good performance for the time, they did not stack up well against the latest A8 SOCs from Apple. The A53 and A57 designs were also announced around two years ago. These are the first 64 bit designs from ARM and were meant to compete with the latest custom designs from Apple and Qualcomm’s upcoming 64 bit part. We are only now just seeing these parts make it into production, and even Qualcomm has licensed the A53 and A57 designs to insure a faster time to market for this latest batch of next-generation mobile devices.
We can look back over the past five years and see that ARM is moving forward in announcing their parts and then having their partners ship them within a much shorter timespan than we were used to seeing. ARM is hoping to accelerate the introduction of its new parts within the next year.
Battlefield 4 Results
At the end of my first Frame Rating evaluation of the GTX 970 after the discovery of the memory architecture issue, I proposed the idea that SLI testing would need to be done to come to a more concrete conclusion on the entire debate. It seems that our readers and the community at large agreed with us in this instance, repeatedly asking for those results in the comments of the story. After spending the better part of a full day running and re-running SLI results on a pair of GeForce GTX 970 and GTX 980 cards, we have the answers you're looking for.
Today's story is going to be short on details and long on data, so if you want the full back story on what is going on why we are taking a specific look at the GTX 970 in this capacity, read here:
- Part 1: NVIDIA issues initial statement
- Part 2: Full GTX 970 memory architecture disclosed
- Part 3: Frame Rating: GTX 970 vs GTX 980
- Part 4: Frame Rating: GTX 970 SLI vs GTX 980 SLI (what you are reading now)
Okay, are we good now? Let's dive into the first set of results in Battlefield 4.
Battlefield 4 Results
Just as I did with the first GTX 970 performance testing article, I tested Battlefield 4 at 3840x2160 (4K) and utilized the game's ability to linearly scale resolution to help me increase GPU memory allocation. In the game settings you can change that scaling option by a percentage: I went from 110% to 150% in 10% increments, increasing the load on the GPU with each step.
Memory allocation between the two SLI configurations was similar, but not as perfectly aligned with each other as we saw with our single GPU testing.
In a couple of cases, at 120% and 130% scaling, the GTX 970 cards in SLI are actually each using more memory than the GTX 980 cards. That difference is only ~100MB but that delta was not present at all in the single GPU testing.
It has been an abnormal week for us here at PC Perspective. Our typical review schedule has pretty much flown out the window, and the past seven days have been filled with learning, researching, retesting, and publishing. That might sound like the norm, but in these cases the process was initiated by tips from our readers. Last Saturday (24 Jan), a few things were brewing:
- Ryan was informed by NVIDIA that the memory layout of the GTX 970 was different than expected.
- The huge (now 168 page) overclock.net forum thread about the Samsung 840 EVO slowdown was once again gaining traction.
- Someone got G-Sync working on a laptop integrated display.
We had to do a bit of triage here of course, as we can only research and write so quickly. Ryan worked the GTX 970 piece as it was the hottest item. I began a few days of research and testing on the 840 EVO slow down issue reappearing on some drives, and we kept tabs on that third thing, which at the time seemed really farfetched. With those two first items taken care of, Ryan shifted his efforts to GTX 970 SLI testing while I shifted my focus to finding out of there was any credence to this G-Sync laptop thing.
A few weeks ago, an ASUS Nordic Support rep inadvertently leaked an interim build of the NVIDIA driver. This was a mobile driver build (version 346.87) focused at their G751 line of laptops. One recipient of this driver link posted it to the ROG forum back on the 20th. A fellow by the name Gamenab, owning the same laptop cited in that thread, presumably stumbled across this driver, tried it out, and was more than likely greeted by this popup after the installation completed:
Now I know what you’re thinking, and it’s probably the same thing anyone would think. How on earth is this possible? To cut a long story short, while the link to the 346.87 driver was removed shortly after being posted to that forum, we managed to get our hands on a copy of it, installed it on the ASUS G751 that we had in for review, and wouldn’t you know it we were greeted by the same popup!
Ok, so it’s a popup, could it be a bug? We checked NVIDIA control panel and the options were consistent with that of a G-Sync connected system. We fired up the pendulum demo and watched the screen carefully, passing the machine around the office to be inspected by all. We then fired up some graphics benchmarks that were well suited to show off the technology (Unigine Heaven, Metro: Last Light, etc), and everything looked great – smooth steady pans with no juddering or tearing to be seen. Ken Addison, our Video Editor and jack of all trades, researched the panel type and found that it was likely capable of 100 Hz refresh. We quickly dug created a custom profile, hit apply, and our 75 Hz G-Sync laptop was instantly transformed into a 100 Hz G-Sync laptop!
Ryan's Note: I think it is important here to point out that we didn't just look at demos and benchmarks for this evaluation but actually looked at real-world gameplay situations. Playing through Metro: Last Light showed very smooth pans and rotation, Assassin's Creed played smoothly as well and flying through Unigine Heaven manually was a great experience. Crysis 3, Battlefield 4, etc. This was NOT just a couple of demos that we ran through - the variable refresh portion of this mobile G-Sync enabled panel was working and working very well.
At this point in our tinkering, we had no idea how or why this was working, but there was no doubt that we were getting a similar experience as we have seen with G-Sync panels. As I digested what was going on, I thought surely this can’t be as good as it seems to be… Let’s find out, shall we?