Subject: General Tech, Motherboards, Processors | September 20, 2014 - 06:51 PM | Scott Michaud
Tagged: xeon, Haswell-EP, ddr4, ddr3, Intel
Well this is interesting and, while not new, is news to me.
The upper-tier Haswell processors ushered DDR4 into the desktops for enthusiasts and servers, but DIMMs are quite expensive and incompatible with the DDR3 sticks that your organization might have been stocking up on. Despite the memory controller being placed on the processor, ASRock has a few motherboards which claim DDR3 support. ASRock, responding to Anandtech's inquiry, confirmed that this is not an error and Intel will launch three SKUs, one eight-core, one ten-core, and one twelve-core, with a DDR3-supporting memory controller.
The three models are:
|E5-2629 v3||E5-2649 v3||E5-2669 v3|
|Cores (Threads)||8 (16)||10 (20)||12 (24)|
|Clock Rate||2.4 GHz||2.3 GHz||2.3 Ghz|
The processors, themselves, might not be cheap or easily attainable, though. There are rumors that Intel will require customers purchase at least a minimum amount. It might not be worth buying these processors unless you have a significant server farm (or similar situation).
Server and Workstation Upgrades
Today, on the eve of the Intel Developer Forum, the company is taking the wraps off its new server and workstation class high performance processors, Xeon E5-2600 v3. Known previously by the code name Haswell-EP, the release marks the entry of the latest microarchitecture from Intel to multi-socket infrastructure. Though we don't have hardware today to offer you in-house benchmarks quite yet, the details Intel shared with me last month in Oregon are simply stunning.
Starting with the E5-2600 v3 processor overview, there are more changes in this product transition than we saw in the move from Sandy Bridge-EP to Ivy Bridge-EP. First and foremost, the v3 Xeons will be available in core counts as high as 18, with HyperThreading allowing for 36 accessible threads in a single CPU socket. A new socket, LGA2011-v3 or R3, allows the Xeon platforms to run a quad-channel DDR4 memory system, very similar to the upgrade we saw with the Haswell-E Core i7-5960X processor we reviewed just last week.
The move to a Haswell-based microarchitecture also means that the Xeon line of processors is getting AVX 2.0, known also as Haswell New Instructions, allowing for 2x the FLOPS per clock per core. It also introduces some interesting changes to Turbo Mode and power delivery we'll discuss in a bit.
Maybe the most interesting architectural change to the Haswell-EP design is per core P-states, allowing each of the up to 18 cores running on a single Xeon processor to run at independent voltages and clocks. This is something that the consumer variants of Haswell do not currently support - every cores is tied to the same P-state. It turns out that when you have up to 18 cores on a single die, this ability is crucial to supporting maximum performance on a wide array of compute workloads and to maintain power efficiency. This is also the first processor to allow independent uncore frequency scaling, giving Intel the ability to improve performance with available headroom even if the CPU cores aren't the bottleneck.
Subject: Motherboards | August 30, 2014 - 01:10 PM | Tim Verry
Tagged: xeon, x99 ws, X99, workstation, socket 2011-3, Intel, Haswell-E, asrock
Alongside the X99 Extreme6 for enthusiasts, ASRock has launched the X99 WS (E-ATX) motherboard for professional workstations and servers. The board uses a black PCB, blue aluminum heatsinks on the power phase and PCH areas, and gold colored caps. The E-ATX motherboard uses short screws and a thinner CPU socket backplate to allow it to fit into 1U server cases. You have ample PCI-E slots, DDR4 (and ECC RDIMMs), USB 3.0, Thunderbolt 2 (with an AIC), eSATA, M.2, and SATA III for connectivity. Power is regulated by a 12 power phase "Digi Power" design with high quality Ultra Dual-N MOSFETs, 60A chokes, 12K Nichicon capacitors.
The X99 WS also uses larger heatsinks, especially around the CPU area. The board features an Intel LGA 2011-3 socket that will accept Haswell-E or Xeon E5-1600/2600 v3 processors up to 18 cores and 160W TDPs. Eight memory slots surround the CPU socket and support a maximum of 128GB of DDR4 3200+ memory. Storage is handled by 10 SATA III 6Gbps ports and a single "Ultra M.2" slot that supports SATA III or PCI-E x4 controller equipped drives.
This worskation board takes advantage of the Extended ATX form factor to cram in additional PCI-E 3.0 slots versus the shorter X99 Extreme6. A PCI-E 2.0 x 16 slot sits directly below the CPU socket followed by the (mentioned above) M.2 slot and five PCI-E 3.0 x 16 slots. The five PCI-E 3.0 slots allow for 4-Way CrossFireX and 4-Way SLI multi-GPU setups. To take full advantage of this board, you will want to pair it with the higher end Haswell-E parts (5930K, 5960X, et al) with the full 40 lanes rather than the entry level 5820K with its limited 28 PCI-E lanes.
Using internal headers, the X99 WS supports a TPM module, two COM ports, two USB 3.0 ports, four USB 2.0 ports, and a Thunderbolt port (via an extra add-in-card). It features two CPU fan and three chassis fan headers as well.
The X99 WS features the following ports on the rear IO panel.
- 1 x PS/2
- 4 x USB 2.0
- 1 x eSATA
- 4 x USB 3.0
- 2 x Gigabit LAN
- 1 x Intel I217LM
- 1 x Intel I210AT
- 7.1 Channel Audio (Realtek ALC1150)
- 5 x Analog audio outputs
- 1 x Optical audio output
The X99 WS is available now with a three year manufaturer warranty for $323.99. The price alone pushes this board well into the professional market, but gamers can get most of the way to the WS feature-wise with the X99 Extreme6. The WS goes up against the boards like the ASUS X99-Deluxe and MSI X99S XPOWER AC.
Subject: General Tech, Graphics Cards, Processors | July 19, 2014 - 03:05 AM | Scott Michaud
Tagged: Xeon Phi, xeon, Intel, avx-512, avx
It is difficult to know what is actually new information in this Intel blog post, but it is interesting none-the-less. Its topic is the AVX-512 extension to x86, designed for Xeon and Xeon Phi processors and co-processors. Basically, last year, Intel announced "Foundation", the minimum support level for AVX-512, as well as Conflict Detection, Exponential and Reciprocal, and Prefetch, which are optional. This, earlier blog post was very much focused on Xeon Phi, but it acknowledged that the instructions will make their way to standard, CPU-like Xeons at around the same time.
This year's blog post brings in a bit more information, especially for common Xeons. While all AVX-512-supporting processors (and co-processors) will support "AVX-512 Foundation", the instruction set extensions are a bit more scattered.
|Conflict Detection Instructions||Yes||Yes||Yes|
|Exponential and Reciprocal Instructions||No||Yes||Yes|
|Byte and Word Instructions||Yes||No||No|
|Doubleword and Quadword Instructions||Yes||No||No|
|Vector Length Extensions||Yes||No||No|
Source: Intel AVX-512 Blog Post (and my understanding thereof).
So why do we care? Simply put: speed. Vectorization, the purpose of AVX-512, has similar benefits to multiple cores. It is not as flexible as having multiple, unique, independent cores, but it is easier to implement (and works just fine with having multiple cores, too). For an example: imagine that you have to multiply two colors together. The direct way to do it is multiply red with red, green with green, blue with blue, and alpha with alpha. AMD's 3DNow! and, later, Intel's SSE included instructions to multiply two, four-component vectors together. This reduces four similar instructions into a single operating between wider registers.
Smart compilers (and programmers, although that is becoming less common as compilers are pretty good, especially when they are not fighting developers) are able to pack seemingly unrelated data together, too, if they undergo similar instructions. AVX-512 allows for sixteen 32-bit pieces of data to be worked on at the same time. If your pixel only has four, single-precision RGBA data values, but you are looping through 2 million pixels, do four pixels at a time (16 components).
For the record, I basically just described "SIMD" (single instruction, multiple data) as a whole.
This theory is part of how GPUs became so powerful at certain tasks. They are capable of pushing a lot of data because they can exploit similarities. If your task is full of similar problems, they can just churn through tonnes of data. CPUs have been doing these tricks, too, just without compromising what they do well.
Subject: General Tech, Graphics Cards, Processors | July 2, 2014 - 03:55 AM | Scott Michaud
Tagged: Intel, Xeon Phi, xeon, silvermont, 14nm
Anandtech has just published a large editorial detailing Intel's Knights Landing. Mostly, it is stuff that we already knew from previous announcements and leaks, such as one by VR-Zone from last November (which we reported on). Officially, few details were given back then, except that it would be available as either a PCIe-based add-in board or as a socketed, bootable, x86-compatible processor based on the Silvermont architecture. Its many cores, threads, and 512 bit registers are each pretty weak, compared to Haswell, for instance, but combine to about 3 TFLOPs of double precision performance.
Not enough graphs. Could use another 256...
The best way to imagine it is running a PC with a modern, Silvermont-based Atom processor -- only with up to 288 processors listed in your Task Manager (72 actual cores with quad HyperThreading).
The main limitation of GPUs (and similar coprocessors), however, is memory bandwidth. GDDR5 is often the main bottleneck of compute performance and just about the first thing to be optimized. To compensate, Intel is packaging up-to 16GB of memory (stacked DRAM) on the chip, itself. This RAM is based on "Hybrid Memory Cube" (HMC), developed by Micron Technology, and supported by the Hybrid Memory Cube Consortium (HMCC). While the actual memory used in Knights Landing is derived from HMC, it uses a proprietary interface that is customized for Knights Landing. Its bandwidth is rated at around 500GB/s. For comparison, the NVIDIA GeForce Titan Black has 336.4GB/s of memory bandwidth.
Intel and Micron have worked together in the past. In 2006, the two companies formed "IM Flash" to produce the NAND flash for Intel and Crucial SSDs. Crucial is Micron's consumer-facing brand.
So the vision for Knights Landing seems to be the bridge between CPU-like architectures and GPU-like ones. For compute tasks, GPUs edge out CPUs by crunching through bundles of similar tasks at the same time, across many (hundreds of, thousands of) computing units. The difference with (at least socketed) Xeon Phi processors is that, unlike most GPUs, Intel does not rely upon APIs, such as OpenCL, and drivers to translate a handful of functions into bundles of GPU-specific machine language. Instead, especially if the Xeon Phi is your system's main processor, it will run standard, x86-based software. The software will just run slowly, unless it is capable of vectorizing itself and splitting across multiple threads. Obviously, OpenCL (and other APIs) would make this parallelization easy, by their host/kernel design, but it is apparently not required.
It is a cool way that Intel arrives at the same goal, based on their background. Especially when you mix-and-match Xeons and Xeon Phis on the same computer, it is a push toward heterogeneous computing -- with a lot of specialized threads backing up a handful of strong ones. I just wonder if providing a more-direct method of programming will really help developers finally adopt massively parallel coding practices.
I mean, without even considering GPU compute, how efficient is most software at splitting into even two threads? Four threads? Eight threads? Can this help drive heterogeneous development? Or will this product simply try to appeal to those who are already considering it?
Subject: General Tech | June 26, 2014 - 02:36 PM | Ken Addison
Tagged: xeon, video, seiki, podcast, nvidia, msi, Intel, HDMI 2.0, gt70 2pe, gt70, gameworks, FX-9590, displayport 1.3, coolermaster, amd, 4k
PC Perspective Podcast #306 - 06/26/2014
Join us this week as we discuss our Budget PC Shootout, the Coolermaster Elite 110, an AMD GameWorks competitor and more!
The URL for the podcast is: http://pcper.com/podcast - Share with your friends!
- iTunes - Subscribe to the podcast directly through the Store
- RSS - Subscribe through your regular RSS reader
- MP3 - Direct download link to the MP3 file
Hosts: Ryan Shrout, Josh Walrath, Jeremy Hellstrom, and Allyn Maleventano
Welcome to our beta podcast!
Quakecon is coming up, as is our Hardware Workshop!
Week in Review:
0:24:00 Noctua NH-U12S CPU Cooler Review
News items of interest:
0:42:00 You got your FPGA in my Xeon!
Hardware/Software Picks of the Week:
Subject: General Tech | June 19, 2014 - 01:19 PM | Jeremy Hellstrom
Tagged: xeon, Intel, FPGA
Intel has just revealed what The Register is aptly referring to as the FrankenChip, a hybrid Xeon E5 and FPGA chip. This will allow large companies to access the power of a Xeon and be able to offload some work onto an FPGA they can program and optimize themselves. The low power FPGA is actually on the chip, as opposed to Microsoft's recent implementation which saw FPGA's added to PCIe slots. Intel's solution does not use up a slot and also offers direct access to the Xeon cache hierarchy and system memory via QPI which will allow for increased performance. Another low power shot has been fired at ARM's attempts to grow their share of the server market but we shall see if the inherent complexity of programming an FPGA to work with an x86 is more or less attractive than switching to ARM.
"Intel has expanded its chip customization business to help it take on the hazy threat posed by some of the world's biggest clouds adopting low-power ARM processors."
Here is some more Tech News from around the web:
- Amazon's new, not-really-3D Fire: Puts Bezos' cash register in YOUR pocket @ The Register
- Amazon Fire Phone will crash and burn @ The Inquirer
- Knitted Circuit Board Lends Flexibility to E-Textiles @ Hack a Day
- 3D Windowing System Developed Using Wayland, Oculus Rift @ Slashdot
- Google Play Store is littered with 'secret keys' @ The Inquirer
- How farsighted is Microsoft's Azure RemoteApp? @ The Register
- Rollei Mini WiFi Camcorder 1 Review @ NikKTech
- The Dell Inspiron 3000 & 5000 Launch Report @ Tech ARP
Subject: General Tech | April 3, 2014 - 03:00 PM | Jeremy Hellstrom
Tagged: adata, ddr4, xeon
ADATA has been rather busy lately, the release of the brand new Premiere Pro SSD family and now the launch of DDR4 modules for the next generation of Xeon processors. These new DIMMs follow the current trend of energy efficiency in the server room by dropping the required voltage to 1.2V which can add up to quite a bit in a large server farm. The specified speed of 2133MHz is attractive for a first gen server RDIMM though there does not seem to be much information available on the timings.
Taipei, Taiwan – April 3, 2014 - ADATA Technology, a leading manufacturer of high-performance DRAM modules and NAND Flash application products, has announced the launch of new DDR4 modules. Working in close cooperation with Intel, ADATA has successfully developed and launched DDR4 RDIMM (ECC Registered DIMM) that are fully compatible with the newly announced, next generation platform of Intel Xeon processor E5-2600 v3 product family.
Coming in densities of 4, 8 & 16 gigabytes, the new modules run at 1.2 volts, and at a frequency of 2133MHz. The higher clock frequencies, faster data transfer rates, and low voltage operation of DDR4 memory make it especially suited for use in the growing cloud server, storage and networking application fields.
According to Jacky Yang, Product Manager at ADATA: “We are enthusiastic about the great potential of this new DDR4 specification, and we will move quickly to bring this new technology to our customers. Currently in development are DDR4 versions of ECC SO-DIMM, VLP RDIMM, & LRDIMM, so we look forward to providing the stability and reliability that ADATA is known for in a low voltage and high performance package.”
Subject: General Tech | September 11, 2013 - 01:30 PM | Jeremy Hellstrom
Tagged: Ivy Bridge-EP, xeon, xeon E5-2600 v2, idf 2013
A second coming of the Xeon E5-2600 family uses the Ivy Bridge-EP architecture and will sport up to 12 cores, using 22nm Trigate technology. The three CPUs which will be arriving are each aimed at a separate market segment with different core counts and TDP. The lower power chips will sport either 4 or 6 cores and have a TDP between 40-80W with the same 15MB L3 cache as SB-EP. The second has a 25MB L3 cache, 6, 8 or 10 cores and TDPs ranging from 70-130W and uses the same interconnects as previously existed. The last is the beast with 12 cores, TDPs of 115-130W and three rings linking the cores and cache segments with a split memory controller. Check The Register for more info on the high powered end of IDF.
"Companies with workloads that like to ride on lots of threads and cores are going to be able to get a lot more bang for a two-socket box thanks to the launch of the "Ivy Bridge-EP" Xeon E5-2600 v2 processors by Intel."
Here is some more Tech News from around the web:
- Intel reveals 14nm PC, declares Moore's Law 'alive and well' @ The Register
- Fujitsu femto boost promises to double LTE speeds @ The Register
Subject: General Tech | July 23, 2013 - 03:44 PM | Jeremy Hellstrom
Tagged: Intel, atom, 14nm, Avoton, Broadwell, Denverton, xeon, rangeley
Intel has spent the day announcing new products for the server room, from new Atoms to Xeons. Atom will bear the names of Avoton and Rangeley, Avoton will deal with microservers where power and heat are a major concern while Rangeley will appear in network devices and possibly mobile communication devices. In the case of Avoton it will be replacing a chip that has not yet been released, the 32nm Atom S1200 lineup is due out in the near future and will fill a new niche for Intel that Centerton failed to fill. The Register talks a bit more indepth here.
Slightly more powerful will be new Broadwell and Denverton Xeons, the first SoC server chips from Intel which will be manufactured on the 14nm process. We heard much less about these upcoming chips, due for 2014 but you can read what is available at The Inquirer.
"SAN FRANCISCO: CHIPMAKER Intel has revealed more details about its server processor roadmap, including its upcoming Atom chips codenamed Avoton and Rangeley and new 14nm Xeon and Atom parts codenamed Broadwell and Denverton, respectively."
Here is some more Tech News from around the web:
- Multiple Raspberry Pi boards used to create video wall @ Hack a Day
- PIN-Cracking Robot To Be Showed Off At Defcon @ Slashdot
- Flexible PCB makers gearing up mass production for new iPhones @ DigiTimes
- ARM servers to gain boost from ARM, Oracle Java partnership @ The Register
- Linksys EA6700 review: dual-core high-end ac router @ Hardware.info
- UK flicks switch on 'I am a pervert' web filters @ The Register
- Win one of Raijintek's new CPU coolers @ Kitguru