Subject: Mobile | October 7, 2014 - 01:26 PM | Jeremy Hellstrom
Tagged: nvidia, msi, maxwell, GTX 980M, GTX 970M, gt dominator, gs stealth, gs ghost
You've heard about NVIDIA's new GTX900M Series and MSI has released two new families of gaming laptops which contain the new GPU. The GS Stealth and Ghost series are the thinner, lighter more mobile of the laptops while the Dominator Series is more a desktop replacement and should also give you a good workout while you are carting them around. The base model will run you $1600, with more expensive options available such as the limited edition Crimson Red Stealth Pro model at the end. Before you ask, no the integral displays are not G-SYNC however since the mobile GPUs are based on Maxwell you may be able to output to a G-SYNC monitor with a variable refresh. Stay tuned for more.
City of Industry, Calif. – October 7, 2014 – MSI Computer Corp, a leading manufacturer of computer hardware products and solutions, announces the immediate availability of the critically acclaimed GS Stealth/Ghost Series and GT Dominator Series gaming laptops with NVIDIA’s latest GTX 900M Series graphics. Armed with unprecedented power and an array of cutting-edge features, including MSI’s new SHIFT technology, the new gaming notebooks deliver up to 28% more graphics performance for a mobile gaming experience without barriers.
“MSI’s newest gaming laptops showcase breakthroughs in mobile gaming technology that improve graphics performance, increase gaming comfort, and transport gamers into an unbelievable gaming journey,” says Andy Tung, president of MSI Pan America. “NVIDIA’s GTX 900M Series GPU delivers mobile graphics that are up to par with desktop graphics card models and superior to anything we’ve ever seen.”
SHIFT Your Speed
MSI’s newest gaming laptops feature their exclusive SHIFT power adjustment technology that enables easy tweaking of CPU and GPU performance to best suit the gamers’ needs. SHIFT comes with three proprietary modes: Sport to maximize CPU and GPU usage for extreme performance, Comfort for a smooth and balanced ride, and Green, which enables the lowest power consumption of both CPU and GPU while maintaining the coolest constant temperatures for both.
MSI provides unprecedented customization in all NVIDIA GeForce GTX 900M Series graphics equipped laptops via the Dragon Gaming Center and SteelSeries Engine. Gamers can SHIFT CPU and GPU usage through the Dragon Gaming Center as well as monitor system performance, temperature, network speed, power consumption, fan speed and more. The SteelSeries Engine gives gamers the ability to personalize playing style with over a billion customization options, program individual keys for unlimited configurations, determine key color and lighting patterns, save and share configurations, and learn gaming patterns with key usage statistics.
Cutting Edge Components
All revamped gaming laptops come with 4th Gen Intel Core i7 processors, Killer E2200 Game Networking, Sound Blaster Cinema, Dynaudio Technology, XSplit Gamecaster, 4K HDMI Output, Matrix Display and NVIDIA Surround View. NVIDIA Surround is now supported on all next gen models, allowing gamers to immersive themselves in the ultimate gaming experience. Select models come with MSI’s Super RAID technology which supports up to 4x M.2 SATA SSD’s in RAID 0, ultra-high resolution 3K displays, and Killer DoubleShot Pro combining Killer E2200 Game Networking with Killer N1525 Wireless AC.
MSI’s latest update applies to the ultra-thin and light GS70 Stealth Pro and GS60 Ghost Pro models, and the potent GT72 Dominator, GT70 Dominator and GT60 Dominator gaming laptops. All GS and GT gaming laptops equipped with NVIDIA’s GTX 900M graphics are available now starting at $1,599.99.
In addition, MSI is launching a special edition GS70 Stealth Pro in Crimson Red, catering to the demands of gamers who want more choices in color and style. The GS70 Stealth Pro Crimson Red edition will be available through online retailers with next gen graphics.
If there is one message that I get from NVIDIA's GeForce GTX 900M-series announcement, it is that laptop gaming is a first-class citizen in their product stack. Before even mentioning the products, the company provided relative performance differences between high-end desktops and laptops. Most of the rest of the slide deck is showing feature-parity with the desktop GTX 900-series, and a discussion about battery life.
First, the parts. Two products have been announced: The GeForce GTX 980M and the GeForce GTX 970M. Both are based on the 28nm Maxwell architecture. In terms of shading performance, the GTX 980M has a theoretical maximum of 3.189 TFLOPs, and the GTX 970M is calculated at 2.365 TFLOPs (at base clock). On the desktop, this is very close to the GeForce GTX 770 and the GeForce GTX 760 Ti, respectively. This metric is most useful when you're compute bandwidth-bound, at high resolution with complex shaders.
The full specifications are:
|GTX 980M||GTX 970M||
|Memory||Up to 4GB||Up to 3GB||4GB||4GB||4GB/8GB|
|Memory Rate||2500 MHz||2500 MHz||7.0 (GT/s)||7.0 (GT/s)||2500 MHz|
As for the features, it should be familiar for those paying attention to both desktop 900-series and the laptop 800M-series product launches. From desktop Maxwell, the 900M-series is getting VXGI, Dynamic Super Resolution, and Multi-Frame Sampled AA (MFAA). From the latest generation of Kepler laptops, the new GPUs are getting an updated BatteryBoost technology. From the rest of the GeForce ecosystem, they will also get GeForce Experience, ShadowPlay, and so forth.
For VXGI, DSR, and MFAA, please see Ryan's discussion for the desktop Maxwell launch. Information about these features is basically identical to what was given in September.
BatteryBoost, on the other hand, is a bit different. NVIDIA claims that the biggest change is just raw performance and efficiency, giving you more headroom to throttle. Perhaps more interesting though, is that GeForce Experience will allow separate one-click optimizations for both plugged-in and battery use cases.
The power efficiency demonstrated with the Maxwell GPU in Ryan's original GeForce GTX 980 and GTX 970 review is even more beneficial for the notebook market where thermal designs are physically constrained. Longer battery life, as well as thinner and lighter gaming notebooks, will see tremendous advantages using a GPU that can run at near peak performance on the maximum power output of an integrated battery. In NVIDIA's presentation, they mention that while notebooks on AC power can use as much as 230 watts of power, batteries tend to peak around 100 watts. Given that a full speed, desktop-class GTX 980 has a TDP of 165 watts, compared to the 250 watts of a Radeon R9 290X, translates into notebook GPU performance that will more closely mirror its desktop brethren.
Of course, you probably will not buy your own laptop GPU; rather, you will be buying devices which integrate these. There are currently five designs across four manufacturers that are revealed (see image above). Three contain the GeForce GTX 980M, one has a GTX 970M, and the other has a pair of GTX 970Ms. Prices and availability are not yet announced.
Subject: General Tech, Mobile | October 4, 2014 - 11:43 PM | Tim Verry
Amazon launched a bevy of new tablets and eReaders late last month. An updated Kindle Fire HDX 8.9 will be available October 21st and replaces last year’s model as the flagship Kindle tablet.
Measuring 8.9” and weighing 13.2 ounces (13.7oz for the 4G model), the upcoming HDX 8.9” tablet retains the same form factor as its predecessor. The mobile device does pack in internal hardware improvements, updated software features, and an optional bluetooth Fire Keyboard accessory. The tablet features an 8.9” 2560x1600 (339 PPI) display, Dobly Atmos audio, an 8MP rear camera, and a 720p front-facing webcam. Amazon is using what it calls “Dynamic Light Control” which alters the display’s color temperature to match ambient light along with a dynamic backlighting.
With that said, the major changes between the previous model and the new Fire HDX 8.9 lie in updated internal hardware and Amazon’s latest operating system and UI features. Specifically, Amazon has upgraded to the Qualcomm Snapdragon 805 SoC and a faster (802.11ac) Wi-Fi radio. Last year’s tablet used the Snapdragon 800 clocked at 2.2GHz with an Adreno 330 GPU. The new HDX 8.9’s Snapdragon 805 is clocked at 2.5GHz and features an Adreno 420 GPU. Both tablets feature 2GB RAM and 16GB, 32GB, and 64GB internal storage options. Amazon claims up to 12 hours of battery life when reading, browsing the web, and watching videos.
The new tablet with ship with Amazon’s latest Fire OS operating system known as Fire OS 4.0 “Sangria.” Sangria is a custom fork of Android 4.4 KitKat that features Amazon’s custom user interface in addition to ASAP predictive pre-loading, Smart Suspend technology, Family Library media sharing, device profiles, and Mayday help/support.
The Fire Keyboard is a 7oz, 4.8mm thin Bluetooth keyboard that can attach magnetically to the optional Origami tablet case. The keyboard features 74 keys (including shortcut and search keys), a small trackpad, Bluetooth 3.0, and a lithium ion battery that has a claimed battery life of two months of active usage. It has an MSRP of $59.99 and will begin shipping October 21.The Origami case is another optional accessory that can hold and protect both the keyboard and tablet. It will be available in black, blue, and red leather for $69.99 from Amazon.
The new Kindle Fire HDX 8.9 is up for pre-order now but will officially be available on October 21st. Pricing and SKU data is listed in the table below. Needless to say, this is an overall minor upgrade over last year's model and it is not something existing HDX 8.9" users are likely excited about. However, on its own, the new model has decent specifications and if you are looking into a new Kindle Fire tablet, it is worth considering, especially if you plan to take advantage of the faster processor for activities such as gaming in addition to reading books and watching movies.
Pricing Information for Amazon's New Kindle Fire HDX 8.9" Tablet
Will you be picking up the latest Fire OS flagship?
Introduction and Design
A little over a year ago, we posted our review of the Lenovo Y500, which was a gaming notebook that leveraged not one, but two discrete video adapters (2 x NVIDIA GeForce GT 650M in SLI, to be exact) to achieve respectable gaming performance at a reasonable price point (around $1,200 at the time of the review).
Well—take away nearly a pound of weight (to 5.7 lbs), slim the case down to around an inch thick, update the chipset, and remove one video card, and you’ve got the Lenovo Y50 Touch, which ought to be able to improve upon the Y500 in nearly every area if the specifications add up to typical results. Here’s the full list of what our review unit includes:
While the GTX 860M (2 GB) is a far cry from, say, the GTX 880M (8 GB) we had the pleasure of testing in MSI’s GT70 2PE, it’s still a very capable card that should provide satisfactory results without breaking the bank (or the back). The rest of the spec sheet is conventional fare for a budget gaming notebook, with the only other surprise being the inclusion of a touchscreen—an option which replaces the traditional matte LCD panel in the standard Y50.
The configuration we received has already been slightly updated to include a CPU that’s a nudge better than the i7-4700HQ: the i7-4710HQ (which gains it 100 MHz in Turbo Boost clock rate). Otherwise, the specs are identical, and the street price is very close to that of the Y500 we originally reviewed: $1,139. Currently, an extra 10 bucks will also score you an external DVD+/-RW drive, and just 90 bucks more will boost your GTX 860M’s VRAM to 4 GB (from 2 GB) and your system RAM to 16 GB from 8 GB. That’s really not a bad deal at all.
Subject: General Tech, Mobile | October 1, 2014 - 03:17 PM | Scott Michaud
Tagged: Kickstarter, Firefox OS, web, chromecast
When Google released the Chromecast, it was a surprisingly clean solution for streaming video (my apologies if solutions existed before it). Just plug it into HDMI and connect to it with a PC or a mobile device to use the TV as monitor for content, and it is cheap. I figured that the open source community would like one of their own, but I did not think it was going to be done. Now there is a Kickstarter up, with FirefoxOS.
I constantly struggle with whether to discuss crowdfunding because, on the one hand, you never know if something will tank. On the other hand, is it really any less sketchy than pre-release information for computer hardware or video games (especially pre-release news for video games)?
In this case, I found out that it was promoted by Mozilla on their Hacks blog. It is based on a Rockchip 3066 SoC with 1GB RAM, 4GB of storage and 2.4 GHz Wireless-N. As stated earlier, it runs FirefoxOS which means that apps are websites. The SoC has a Mali-400 GPU that is capable of OpenGL ES 2.0, so it might even be able to support WebGL if the software and drivers are certified. Don't expect jaw-dropping 3D graphics, though. The GPU is rated at about 9 GFLOPs. For comparison, the Tegra K1 has a peak compute throughput of about 365 GFLOPs; alternatively, it is fairly close to later-model Intel GMA graphics (not Intel HD Graphics... GMA). Still, it might allow for some interesting 2D (or simplistic 3D) games.
Just a day-or-so in, it is already at over 150% funding.
Subject: Systems, Mobile | September 30, 2014 - 04:15 AM | Scott Michaud
Tagged: Windows 8.1, hp, cheap tablet, cheap computer
Before I get into the devices, the $149 HP Stream 8 tablet and certain models of the HP Stream 13 laptop (the ones with an optional 4G modem) includes "free 4G for life" for customers in the USA. Reading in the fine print, the device company apparently signed a deal with T-Mobile for 200MB/mo of 4G service. Of course, 200MB will barely cover the Windows Update regimen of certain months, but you have WiFi for that. It is free, and free is good. I can guess that T-Mobile is crossing their fingers that dripping a drop of water on the tongues of the thirsty will convince them to go to the fountain.
If it works? Great. That is just about the most honest way that I have ever seen a telecom company attract new customers.
Back to these devices. Oh right, they're cheap. They are so cheap, they barely have any technical specifications. The $199.99 HP Stream 11 laptop has an 11-inch display. The $229.99 HP Stream 13 laptop has a 13-inch display and can be configured with an optional 4G modem. Both are passively cooled (more fanless PCs...) and run on a dual-core processor. Both provide a year of Office 365 Personal subscriptions. Both are available in blueish-purple or pinkish-purple.
The two tablets (7-inch Stream 7 and 8-inch Stream 8) are a similar story. They run an x86 processor with full Windows 8.1 and a year's subscription to Office 365. Somehow, the tablets are based on Intel quad-core CPUs (rather than the laptop's passively cooled dual-cores) despite being cheaper. Then again, they could be completely different architectures.
While HP is interested in, you know, selling product, I expect that Microsoft's generous licensing terms (see also the Toshiba alternative we reported earlier) is an attempt to push their cloud services. They know that cheaper device categories cannot bare as much royalties as a fully-featured laptop, and not having a presence at those prices is conceding it to Google -- and conceding that to Google is really giving up on cloud services for those customers. The simple solution? Don't forfeit those markets, just monetize with your own cloud service. I doubt that it will harm their higher-end devices.
One Small Step
While most articles surrounding the iPhone 6 and iPhone 6 Plus this far have focused around user experience and larger screen sizes, performance, and in particular the effect of Apple's transition to the 20nm process node for the A8 SoC have been our main questions regarding these new phones. Naturally, I decided to put my personal iPhone 6 though our usual round of benchmarks.
First, let's start with 3DMark.
Comparing the 3DMark scores of the new Apple A8 to even the last generation A7 provides a smaller improvement than we are used to seeing generation-to-generation with Apple's custom ARM implementations. When you compare the A8 to something like the NVIDIA Tegra K1, which utilizes desktop-class GPU cores, the overall score blows Apple out of the water. Even taking a look at the CPU-bound physics score, the K1 is still a winner.
A 78% performance advantage in overall score when compared the A8 shows just how much of a powerhouse NVIDIA has with the K1. (Though clearly power envelopes are another matter entirely.)
If we look at more CPU benchmarks, like the browser-based Google Octane and SunSpider tests, the A8 starts to shine more.
While the A8 edges out the A7 to be the best performing device and 54% faster than the K1 in SunSpider, the A8 and K1 are neck and neck in the Google Octane benchmark.
Moving back to a graphics heavy benchmark, GFXBench's Manhattan test, the Tegra K1 has a 75% percent performance advantage over the A8 though it is 36% faster than the previous A7 silicon.
These early results are certainly a disappointment compared to the usual generation-to-generation performance increase we see with Apple SoCs.
However, the other aspect to look at is power efficiency. With normal use I have noticed a substantial increase in battery life of my iPhone 6 over the last generation iPhone 5S. While this may be due to a small (about 1 wH) increase in battery capacity, I think more can be credited to this being an overall more efficient device. Certain choices like sticking to a highly optimized Dual Core CPU design and Quad Core GPU, as well as a reduction in process node to 20nm all contribute to increased battery life, while surpassing the performance of the last generation Apple A7.
In that way, the A8 moves the bar forward for Apple and is a solid first attempt at using the 20nm silicon technology at TSMC. There is a strong potential that further refined parts (like the expected A8x for the iPad revisions) Apple will be able to further surpass 28nm silicon in performance and efficiency.
Subject: Graphics Cards, Processors, Mobile | September 29, 2014 - 01:53 AM | Scott Michaud
Tagged: apple, a8, a7, Imagination Technologies, PowerVR
First, Chipworks released a dieshot of the new Apple A8 SoC (stored at archive.org). It is based on the 20nm fabrication process from TSMC, which they allegedly bought the entire capacity for. From there, a bit of a debate arose regarding what each group of transistors represented. All sources claim that it is based around a dual-core CPU, but the GPU is a bit polarizing.
Image Credit: Chipworks via Ars Technica
Most sources, including Chipworks, Ars Technica, Anandtech, and so forth believe that it is a quad-core graphics processor from Imagination Technologies. Specifically, they expect that it is the GX6450 from the PowerVR Series 6XT. This is a narrow upgrade over the G6430 found in the Apple A7 processor, which is in line with the initial benchmarks that we saw (and not in line with the 50% GPU performance increase that Apple claims). For programmability, the GX6450 is equivalent to a DirectX 10-level feature set, unless it was extended by Apple, which I doubt.
Image Source: DailyTech
DailyTech has their own theory, suggesting that it is a GX6650 that is horizontally-aligned. From my observation, their "Cluster 2" and "Cluster 5" do not look identical at all to the other four, so I doubt their claims. I expect that they heard Apple's 50% claims, expected six GPU cores as the rumors originally indicated, and saw cores that were not there.
Which brings us back to the question of, "So what is the 50% increase in performance that Apple claims?" Unless they had a significant increase in clock rate, I still wonder if Apple is claiming that their increase in graphics performance will come from the Metal API even though it is not exclusive to new hardware.
But from everything we saw so far, it is just a handful of percent better.
Subject: General Tech, Processors, Mobile | September 27, 2014 - 02:38 PM | Scott Michaud
Tagged: Intel, spreadtrum, rda, Rockchip, SoC
A few months ago, Intel partnered with Rockchip to develop low-cost SoCs for Android. The companies would work together on a design that could be fabricated at TSMC. This time Intel is partnering with Tsinghua Unigroup Ltd. and, unlike Rockchip, also investing in them. The deal will be up to $1.5 billion USD in exchange for a 20% share (approximately) of a division of Tsinghua.
Image Credit: Wikipedia
Intel is hoping to use this partnership to develop mobile SoCs, for smart (and "feature") phones, tablets, and other devices, and get significant presence in the Chinese mobile market. Tsinghua acquired Spreadtrum Communications and RDA Microelectronics within the last two years. The "holding group" that owns these division is apparently the part of Tsinghua which Intel is investing in, specifically.
Spreadtrum will produce SoCs based on Intel's "Intel Architecture". This sounds like they are referring to the 32-bit IA-32, which means that Spreadtrum would be developing 32-bit SoCs, but it is possible that they could be talking about Intel 64. These products are expected for 2H'15.
Subject: General Tech, Mobile | September 26, 2014 - 01:45 AM | Scott Michaud
Tagged: tablet, Nexus, google, nexus 9, nvidia, tegra k1
The Nexus line is due for an update, with each product being released for at least a year. They are devices which embody Google's vision... for their own platform. You can fall on either side of that debate, whether it guides OEM partners or if it is simply a shard the fragmentation issue, if you even believe that fragmentation is bad, but they are easy to recommend and a good benchmark for Android.
We are expecting a few new entries in the coming months, one of which being the Nexus 9. Of note, it is expected to mark the return of HTC to the Nexus brand. They were the launch partner with the Nexus One and then promptly exited stage left as LG, Samsung, and ASUS performed the main acts.
We found this out because NVIDIA spilled the beans on their lawsuit filing against Qualcomm and Samsung. Apparently, "the HTC Nexus 9, expected in the third quarter of 2014, is also expected to use the Tegra K1". It has since been revised to remove the reference. While the K1 has a significant GPU to back it up, it will likely be driving a very high resolution display. The Nexus 6 is expected to launch at around the same time, along with Android 5.0 itself, and the 5.2-inch phone is rumored to have a 1440p display. It seems unlikely that a larger, tablet display will be lower resolution than the phone it launches alongside -- and there's not much room above it.
The Google Nexus 9 is expected for "Q3".