It wouldn’t be February if we didn’t hear the Q4 FY14 earnings from NVIDIA! NVIDIA does have a slightly odd way of expressing their quarters, but in the end it is all semantics. They are not in fact living in the future, but I bet their product managers wish they could peer into the actual Q4 2014. No, the whole FY14 thing relates back to when they made their IPO and how they started reporting. To us mere mortals, Q4 FY14 actually represents Q4 2013. Clear as mud? Lord love the Securities and Exchange Commission and their rules.
The past quarter was a pretty good one for NVIDIA. They came away with $1.144 billion in gross revenue and had a GAAP net income of $147 million. This beat the Street’s estimate by a pretty large margin. As a response, trading of NVIDIA’s stock has gone up in after hours. This has certainly been a trying year for NVIDIA and the PC market in general, but they seem to have come out on top.
NVIDIA beat estimates primarily on the strength of the PC graphics division. Many were focusing on the apparent decline of the PC market and assumed that NVIDIA would be dragged down by lower shipments. On the contrary, it seems as though the gaming market and add-in sales on the PC helped to solidify NVIDIA’s quarter. We can look at a number of factors that likely contributed to this uptick for NVIDIA.
Subject: Graphics Cards | January 21, 2014 - 08:49 PM | Ryan Shrout
Tagged: rumor, nvidia, kepler, gtx titan black, gtx titan, gtx 790
How about some fresh graphics card rumors for your Tuesday afternoon? The folks at VideoCardz.com have collected some information about two potential NVIDIA GeForce cards that are going to hit your pocketbook hard. If the mid-range GPU market was crowded wait until you see the changes NVIDIA might have for you soon on the high-end.
First up is the NVIDIA GeForce GTX TITAN Black Edition, a card that will actually have the same specifications as the GTX 780 Ti but with full performance double precision floating point and a move from 3GB to 6GB of memory. The all-black version of the GeForce GTX 700-series cooler is particularly awesome looking.
Image from VideoCardz.com
The new TITAN would sport the same GPU as GTX 780 Ti, only TITAN BLACK would have higher double precision computing performance, thus more FP64 CUDA cores. The GTX TITAN Black Edition is also said to feature 6GB memory buffer.This is twice as much as GTX 780 Ti, and it pretty much confirms we won’t be seeing any 6GB Ti’s.
The rest is pretty much well known, TITAN BLACK has 2880 CUDA cores, 240 TMUs and 48 ROPs.
VideoCardz.com says this will come in at $999. If true, this is a pure HPC play as the GTX 780 Ti would still offer the same gaming performance for enthusiasts.
Secondly, there looks to be an upcoming dual-GPU graphics card using a pair of GK110 GPUs that will be called the GeForce GTX 790. The specifications that VideoCardz.com says they have indicate that each GPU will have 2496 enabled CUDA cores and a smaller 320-bit memory interface with 5GB designated for each GPU. Cutting back on the memory interface, shader counts and even clocks speeds would allow NVIDIA to manage power consumption at the targeted 300 watt level.
Image from VideoCardz.com
Head over to VideoCardz.com for more information about these rumors but if all goes as they expect, you'll hear about these products quite a bit more in February and March.
What do you think? Are these new $1000 graphics cards something you are looking forward to?
Subject: General Tech | January 6, 2014 - 07:08 PM | Jeremy Hellstrom
Tagged: tegra k1, tegra, SoC, nvidia, kepler, k1, cortex a15, CES, arm, A15
Project X Logan K1 is the first big news out of CES from NVIDIA and represents a bit of a change from what we were expecting. The current belief was that the SoC would have four 28nm Cortex A15 processors but that will only be one flavour of K1, a Denver based dual core version will also be released. Those ARMv8 64-bit processors will natively handle 64 bit applications while the A15 version that The Tech Report had taken pictures of will be limited to 32 bit applications, though that will not matter in many mobile applications. You should also check out Ryan's deep dive into the new Denver and Kepler version here.
"In early 2011, during a CES press event, Nvidia revealed its Project Denver CPU initiative. On Sunday evening, at another CES press conference, the company provided a glimpse of the first Denver-based processor: the Tegra K1. This next-generation SoC features dual Denver CPU cores clocked at up to 2.5GHz. The cores were designed by Nvidia, and they're compatible with the 64-bit ARMv8 instruction set. They have a seven-way superscalar pipeline and a hefty 192KB of L1 cache."
Here is some more Tech News from around the web:
- A Look at NVIDIA’s Most Ambitious SoC Yet: Tegra K1 @ Techgage
- Yahoo adverts infect 27,000 people per hour with malware @ The Inquirer
- Acer cozies up to Google with new 'droid PCs and fondleslab, Chromebook @ The Register
- Trewgrip mobile QWERTY hands-on @ The Inquirer
- China touch panel makers focus on bezel-free applications in 2014 @ DigiTimes
- Intel Ivy Bridge Linux Performance Stepped Up In 2013 @ Phoronix
- AT&T takes aim at T-Mobile with $450 cashback lure @ The Register
- HP: We're axing 29,000 workers? Add another 5,000 to that @ The Register
- Zetta Z15 HD Security Camcorder @ NikKTech
- Seagate updates Backup Plus Slim HDD with smartphone support @ The Inquirer
- MAKE at CES: Follow our Live Blogging Here
Follow all of our coverage of the show at http://pcper.com/ces!
Once known as Logan, now known as K1
NVIDIA has bet big on Tegra. Since the introduction of the SoC's first iteration, that much was clear. With the industry push to mobile computing and the decreased importance of the classic PC design, developing and gaining traction with a mobile processor was not only an expansion of the company’s portfolio but a critical shift in the mindset of a graphics giant.
The problem thus far is that while NVIDIA continues to enjoy success in the markets of workstation and consumer discrete graphics, the Tegra line of silicon-on-chip processors has faltered. Design wins have been tough to come by. Other companies with feet already firmly planted on this side of the hardware fence continue to innovate and seal deals with customers. Qualcomm is the dominant player for mobile processors with Samsung, MediaTek, and others all fighting for the same customers NVIDIA needs. While press conferences and releases have been all smiles and sunshine since day one, the truth is that Tegra hasn’t grown at the rate NVIDIA had hoped.
Solid products based on NVIDIA Tegra processors have been released. The first Google Nexus 7 used the Tegra 3 processor, and was considered the best Android tablet on the market by most, until it was succeeded by the 2013 iteration of the Nexus 7 this year. Tegra 4 slipped backwards, though – the NVIDIA SHIELD mobile gaming device was the answer for a company eager to show the market they built compelling and relevant hardware. It has only partially succeeded in that task.
With today’s announcement of the Tegra K1, previously known as Logan or Tegra 5, NVIDIA hopes to once again spark a fire under partners and developers, showing them that NVIDIA’s dominance in the graphics fields of the PC has clear benefits to the mobile segment as well. During a meeting with NVIDIA about Tegra K1, Dan Vivoli, Senior VP of marketing and a 16 year employee, equated the release of the K1 to the original GeForce GPU. That is a lofty ambition and puts of a lot pressure on the entire Tegra team, not to mention the K1 product itself, to live up to.
Tegra K1 Overview
What we previously knew as Logan or Tegra 5 (and actually it was called Tegra 5 until just a couple of days ago), is now being released as the Tegra K1. The ‘K’ designation indicated the graphics architecture that powers the SoC, in this case Kepler. Also, it’s the first one. So, K1.
The processor of the Tegra K1 look very familiar and include four ARM Cortex-A15 “r3” cores and 2MB of L2 cache with a fifth A15 core used for lower power situations. This 4+1 design is the same that was introduced with the Tegra 4 processor last year and allows NVIDIA to implement a style of “big.LITTLE” design that is unique. Some slight modifications to the cores are included with Tegra K1 that improve performance and efficiency, but not by much – the main CPU is very similar to the Tegra 4.
NVIDIA also unveiled late last night that another version of the Tegra K1 that replaces the quad A15 cores with two of the company's custom designs Denver CPU cores. Project Denver, announced in early 2011, is NVIDIA's attempt at building its own core design based on the ARMv8 64-bit ISA. This puts this iteration of Tegra K1 on the same level as Apple's A7 and Qualcomm's Krait processors. When these are finally available in the wild it will be incredibly intriguing to see how well NVIDIA's architects did in the first true CPU design from the GPU giant.
Follow all of our coverage of the show at http://pcper.com/ces!
Subject: Processors, Mobile | January 6, 2014 - 04:43 AM | Ryan Shrout
Tagged: tegra k1, tegra, SoC, nvidia, kepler, CES 2014, CES
Update: Check out our more in-depth analysis of the Tegra K1 processor from NVIDIA.
Today during its CES 2014 press conference, NVIDIA announced the Tegra K1 SoC as the successor to the Tegra 4 processor. This new ARM-based part includes 192 Kepler-based CUDA cores, sharing the same GPU architecture as the current GeForce GTX 700-series discrete graphics cards.
NVIDIA also announced the Epic has Unreal Engine 4 up and running on the Tegra K1, bringing an entirely new class of games to mobile Android devices. We got to see some demonstrations from NVIDIA running on the K1 and I must admit the visuals were stunning. Frame rates did get a bit choppy during the subway demo of UE4 but it's still early.
As an added surprise, NVIDIA is announcing a version of Tegra K1 that ships with the same quad-core A15 (4+1) design as the Tegra 4 BUT ALSO have a version that uses two NVIDIA Denver CPU cores!! Denver is NVIDIA's custom CPU design based on the ARMv8 architecture, adding 64-bit support to another ARM partner's portfolio.
Tegra K1 is offered in two pin-to-pin compatible versions - a 32-bit quad-core (4-Plus-1 ARM Cortex-A15 CPU) and a custom, NVIDIA-designed 64-bit dual Super Core CPU. This CPU (codenamed “Project Denver”) delivers very high single-thread and multi-thread performance. Both versions deliver stunning graphics and visual computing capabilities powered by the 192-core NVIDIA Kepler GPU.
NVIDIA has only had Denver back for a few days from the fab but there able to showcase it running Android. It's been a long time since the initial announcement of this project and its great to finally see a result.
Tegra K1 with quad-core A15 processor
We'll have an in-depth story on the Tegra K1 on Monday morning, 6am PST right here on PC Perspective so check back then!!
Follow all of our coverage of the show at http://pcper.com/ces!
EVGA Brings Custom GTX 780 Ti Early
Reference cards for new graphics card releases are very important for a number of reasons. Most importantly, these are the cards presented to the media and reviewers that judge the value and performance of these cards out of the gate. These various articles are generally used by readers and enthusiasts to make purchasing decisions, and if first impressions are not good, it can spell trouble. Also, reference cards tend to be the first cards sold in the market (see the recent Radeon R9 290/290X launch) and early adopters get the same technology in their hands; again the impressions reference cards leave will live in forums for eternity.
All that being said, retail cards are where partners can differentiate and keep the various GPUs relevant for some time to come. EVGA is probably the most well known NVIDIA partner and is clearly their biggest outlet for sales. The ACX cooler is one we saw popularized with the first GTX 700-series cards and the company has quickly adopted it to the GTX 780 Ti, released by NVIDIA just last week.
I would normally have a full review for you as soon as we could but thanks to a couple of upcoming trips that will keep me away from the GPU test bed, that will take a little while longer. However, I thought a quick preview was in order to show off the specifications and performance of the EVGA GTX 780 Ti ACX.
As expected, the EVGA ACX design of the GTX 780 Ti is overclocked. While the reference card runs at a base clock of 875 MHz and a typical boost clock of 928 MHz, this retail model has a base clock of 1006 MHz and a boost clock of 1072 MHz. This means that all 2,880 CUDA cores are going to run somewhere around 15% faster on the EVGA ACX model than the reference GTX 780 Ti SKUs.
We should note that though the cooler is custom built by EVGA, the PCB design of this GTX 780 Ti card remains the same as the reference models.
Subject: Graphics Cards | November 8, 2013 - 09:41 PM | Jeremy Hellstrom
Tagged: nvidia, kepler, gtx 780 ti, gk110, geforce
Here is a roundup of the reviews of what is now the fastest single GPU card on the planet, the GTX 780 Ti, which is a fully active GK110 chip. The 7GHz GDDR5 is faster than AMD's memory but use a 384-bit memory bus which is less than the R9 290X which leads to some interesting questions about the performance of this card under high resolutions. Are you willing to pay quite a bit more for better performance and a quieter card? Check out the performance deltas at [H]ard|OCP and see if that changes your mind at all.
You can see how it measures up in ISUs in Ryan's review as well.
"NVIDIA's fastest single-GPU video card is being launched today. With the full potential of the Kepler architecture and GK110 GPU fully unlocked, how will it perform compared to the new R9 290X with new drivers? Will the price versus performance make sense? Will it out perform a TITAN? We find out all this and more."
Here are some more Graphics Card articles from around the web:
- Nvidia's GeForce GTX 780 Ti @ The Tech Report
- Nvidia GeForce GTX 780 Ti @ Bjorn3D
- NVIDIA GeForce GTX 780Ti Review @HiTech Legion
- NVIDIA GeForce GTX 780 Ti 3GB Video Card Review @ Legit Reviews
- NVIDIA GTX 780 Ti Video Card Review @ Hardware Asylum
- NVIDIA GeForce GTX 780 Ti Review @ OCC
- NVIDIA GeForce GTX 780 Ti Performance Review @ Hardware Canucks
- Nvidia GeForce GTX 780 Ti review: Titan killer @ Hardware.info
- Gigabyte GTX 780 GHz Edition 3GB @ eTeknix
- Nvidia GTX780Ti @ Kitguru
- Nvidia GTX 780 Ti @ LanOC Reviews
- NVIDIA GeForce GTX 780 Ti 3 GB @ techPowerUp
- NVIDIA GeForce GTX 780 Ti @ Benchmark Reviews
- Nvidia GTX 780 Ti 3GB @ eTeknix
- Gigabyte Radeon R9 280X @ Legion Hardware
- AMD R9 290 4GB @ eTeknix
GK110 in all its glory
I bet you didn't realize that October and November were going to become the onslaught of graphics cards it has been. I know I did not and I tend to have a better background on these things than most of our readers. Starting with the release of the AMD Radeon R9 280X, 270X and R7 260X in the first week of October, it has pretty much been a non-stop battle between NVIDIA and AMD for the hearts, minds, and wallets of PC gamers.
Shortly after the Tahiti refresh came NVIDIA's move into display technology with G-Sync, a variable refresh rate feature that will work with upcoming monitors from ASUS and others as long as you have a GeForce Kepler GPU. The technology was damned impressive, but I am still waiting for NVIDIA to send over some panels for extended testing.
Later in October we were hit with the R9 290X, the Hawaii GPU that brought AMD back in the world of ultra-class single GPU card performance. It has produced stellar benchmarks and undercut the prices (then at least) of the GTX 780 and GTX TITAN. We tested it in both single and multi-GPU configurations and found that AMD had made some impressive progress in fixing its frame pacing issues, even with Eyefinity and 4K tiled displays.
NVIDIA dropped a driver release with ShadowPlay that allows gamers to record playback locally without a hit on performance. I posted a roundup of R9 280X cards which showed alternative coolers and performance ranges. We investigated the R9 290X Hawaii GPU and the claims that performance is variable and configurable based on fan speeds. Finally, the R9 290 (non-X model) was released this week to more fanfare than the 290X thanks to its nearly identical performance and $399 price tag.
And today, yet another release. NVIDIA's GeForce GTX 780 Ti takes the performance of the GK110 and fully unlocks it. The GTX TITAN uses one fewer SMX and the GTX 780 has three fewer SMX units so you can expect the GTX 780 Ti to, at the very least, become the fastest NVIDIA GPU available. But can it hold its lead over the R9 290X and validate its $699 price tag?
Subject: Graphics Cards | October 28, 2013 - 01:29 PM | Ryan Shrout
Tagged: nvidia, kepler, gtx 780 ti, gtx 780, gtx 770, geforce
A lot of news coming from the NVIDIA camp today, including some price drops and price announcements.
First up, the high-powered GeForce GTX 780 is getting dropped from $649 to $499, a $150 savings that will bring the GTX 780 into line with the competition of AMD's new Radeon R9 290X launched last week.
Next, the GeForce GTX 770 2GB is going to drop from $399 to $329 to help it compete more closely with the R9 280X.
Even you weren't excited about the R9 290X, you have to be excited by competition.
In a surprising turn of events, NVIDIA is now the company with the great bundle deal with GPUs as well! Starting today you'll be able to get a free copy of Batman: Arkham Origins, Splinter Cell: Blacklist and Assassin's Creed IV: Black Flag with the GeForce GTX 780 Ti, GTX 780 and GTX 770. If you step down to the GTX 760 or 660 you'll lose out on the Batman title.
SHIELD discounts are available as well: $100 off you buy the upper tier GPUs and $50 off if you but the lower tier.
UPDATE: NVIDIA just released a new version of GeForce Experience that enabled ShadowPlay, the ability to use Kepler GPUs to record game play in the background with almost no CPU/system ovehead. You can see Scott's initial impressions of the software right here; it seems like its going to be a pretty awesome feature.
Need more news? The yet-to-be-released GeForce GTX 780 Ti is also getting a price - $699 based on the email we just received. And it will be available starting November 7th!!
With all of this news, how does it change our stance on the graphics market? Quite a bit in fact. The huge price drop on the GTX 780, coupled with the 3-game bundle means that NVIDIA is likely offering the better hardware/software combo for gamers this fall. Yes, the R9 290X is likely still a step faster, but now you can get the GTX 780, three great games and spend $50 less.
The GTX 770 is now poised to make a case for itself against the R9 280X as well with its $70 drop. The R9 280X / HD 7970 GHz Edition was definitely a better option with its $100 price delta but with only $30 separating the two competing cards, and the three free games, again the advantage will likely fall to NVIDIA.
Finally, the price point of the GTX 780 Ti is interesting - if NVIDIA is smart they are pricing it based on comparable performance to the R9 290X from AMD. If that is the case, then we can guess the GTX 780 Ti will be a bit faster than the Hawaii card, while likely being quieter and using less power too. Oh, and again, the three game bundle.
NVIDIA did NOT announce a GTX TITAN price drop which might surprise some people. I think the answer as to why will be addressed with the launch of the GTX 780 Ti next month but from what I was hearing over the last couple of weeks NVIDIA can't make the cards fast enough to satisfy demand so reducing margin there just didn't make sense.
NVIDIA has taken a surprisingly aggressive stance here in the discrete GPU market. The need to address and silence critics that think the GeForce brand is being damaged by the AMD console wins is obviously potent inside the company. The good news for us though, and the gaming community as a whole, is that just means better products and better value for graphics card purchases this holiday.
NVIDIA says these price drops will be live by tomorrow. Enjoy!
- NVIDIA GeForce GTX 780 (Newegg.com)
- NVIDIA GeForce GTX 770 (Newegg.com)
- AMD Radeon R9 290X (Newegg.com)
- AMD Radeon R9 280X (Newegg.com)
ShadowPlay is NVIDIA's latest addition to their GeForce Experience platform. This feature allows their GPUs, starting with Kepler, to record game footage either locally or stream it online through Twitch.tv (in a later update). It requires Kepler GPUs because it is accelerated by that hardware. The goal is to constantly record game footage without any noticeable impact to performance; that way, the player can keep it running forever and have the opportunity to save moments after they happen.
Also, it is free.
I know that I have several gaming memories which come unannounced and leave undocumented. A solution like this is very exciting to me. Of course a feature on paper not the same as functional software in the real world. Thankfully, at least in my limited usage, ShadowPlay mostly lives up to its claims. I do not feel its impact on gaming performance. I am comfortable leaving it on at all times. There are issues, however, that I will get to soon.
This first impression is based on my main system running the 331.65 (Beta) GeForce drivers recommended for ShadowPlay.
- Intel Core i7-3770, 3.4 GHz
- NVIDIA GeForce GTX 670
- 16 GB DDR3 RAM
- Windows 7 Professional
- 1920 x 1080 @ 120Hz.
- 3 TB USB3.0 HDD (~50MB/s file clone).
The two games tested are Starcraft II: Heart of the Swarm and Battlefield 3.