All | Editorial | General Tech | Graphics Cards | Networking | Motherboards | Cases and Cooling | Processors | Chipsets | Memory | Displays | Systems | Storage | Mobile | Shows and Expos
Subject: Editorial, General Tech, Graphics Cards | July 17, 2011 - 01:07 PM | Scott Michaud
Tagged: stanford, nvidia, CUDA
NVIDIA has been pushing their CUDA platform for years now as a method to access your GPU for purposes far beyond the scopes of flags and frags. We have seen what a good amount of heterogeneous hardware will do to a process with a hefty portion of parallelizable code from encryption to generating bitcoins; media processing to blurring the line between real-time and non-real-time 3d rendering. NVIDIA also recognizes the role that academia plays in training the future programmers and thus strongly supports when an institution teaches how to use GPU hardware effectively, especially when they teach how to use NVIDIA GPU hardware effectively. Recently, NVIDIA knighted Stanford as the latest of its CUDA Center of Excellence round table.
It will be 150$ if you want it framed.
The list of CUDA Centres of Excellence now currently includes: Georgia Institute of Technology, Harvard School of Engineering, Institute of Process Engineering at Chinese Academy of Sciences, National Taiwan University, Stanford Engineering, TokyoTech, Tsinghua University, University of Cambridge, University of Illinois at Urbana-Champaign, University of Maryland, University of Tennessee, and the University of Utah. If you are interested in learning about programming for GPUs then NVIDIA has just graced blessing on one further choice. Whether that will affect many prospective students and faculty is yet to be seen, but it makes for many amusing puns nonetheless.
Subject: Editorial, Shows and Expos | July 16, 2011 - 12:51 PM | Ryan Shrout
Tagged: sli, sapphire, pdxlan, pdx, nvidia, msi, amd
If you are a PC gaming and live near or around the Portland, OR area you are familiar with the concept of PDXLAN - one of the most popular (but still cool and underground) LAN events in the country. The primary event is going on this weekend and I am here to both game and take a look at what the sponsors are showing off.
MSI has a lot of stuff going on including a look at the latest version of the Afterburner overclocking tool, the 3GB version of the GTX 580 Lightning (that Josh is currently working on a review of) and even a NVIDIA Surround based Dirt 3 sim seat.
Gaming laptops are still taking off here in the US and MSI has a couple on display including a HUGE 18-in mode (on the right) with a keyboard that lights up with various colors of LEDs, configurable.
The Sapphire guys are here as well and are showing off much of what AMD for gamers including Eyefinity configurations like the very popular 5x1 portrait mode. This is something that only AMD offers currently and in this demo we were looking at Dragon Age II. It was definitely grabbing some attention!
Showing that AMD's HD3D technology does indeed have legs Sapphire was showing off the new Samsung SyncMaster SA950 that has a nice external design. I am going to spend some more time with it today to see how it performs, so check back for more!
If you are here, you can also find me getting butt kicked at various games. This is the machine I'll be on, a Maingear built GTX 580 SLI right with an overclocked Intel Core i7 Sandy Bridge processor and 30-in display. I know, it sucks to be me, but someone has to sacrifice and play on it, right?
More from PDXLAN later today!
Subject: Editorial | July 15, 2011 - 06:18 PM | Jeremy Hellstrom
Tagged: killer frogs, friday, folding@home, folding frogs, boinc, 1 billion
When visiting our Forums you will be greeted with an ever changing mix of the old and new. Some members will be tracking down manuals from old kit that they inherited from friends, such as an Epox EP-8KDA3J, while others might have the manual but are having a tough time finding drivers for their DFI LanParty NForce2 Ultra B board. Those threads are nestled between another member's indepth look at the brand new Intel Core i7-2600K with GIGABYTE Z68X-UD7, with plenty of pictures and tips on getting the best overclock.
Our podcast, PC Perspective Podcast #162 this week, has inspired one member to stop lurking and ask about a case for a NAS they want to build, while other long term members discuss the merits of SLI. One member is being very mean to their speakers at Audio Corner while in the Storage Forum you can read about the ressurection of Intel SSDs that suddenly become 8MB in size.
Also this week there has been a lot of talk about Bitcoins, both in terms of the best GPUs to use for mining as well as a rough guide on how much your electricity bill will go up thanks to your new habit, but there are other things you can do with your spare cycles. The PC Perspective Folding Frogs are going strong and helping with protein folding, but it is the PC Perspective Killer Frogs BOINC team which has something to cheer about right now. Our combined BOINC project processed workunit total has gone over 1 billion! Congrats to everyone, old members and new, that have contributed to help us pass this milestone.
Subject: Editorial, General Tech, Systems | July 11, 2011 - 05:57 PM | Scott Michaud
Tagged: xbox, pc gaming
Last week we reported on Microsoft rolling their Games for Windows initiative into Xbox.com and I essentially said that unless Microsoft is trying to roll their now established Xbox brand into Windows that they are missing the point of PC gaming. This week we hear rumors that, in fact, Microsoft may be trying to roll their now established Xbox brand into Windows. According to Insideris, Windows 8 will allow you to play Xbox 360 games on your PC. That said, despite speculation as a result of this news, it does not state whether it will be the complete catalog or a subset of 360 games that are compatible with the PC.
Which came first? The console or the Newegg?
What does this mean for PC gaming? I am unsure at this point. A reasonable outcome would be that Xbox becomes a user-friendly brand for Microsoft’s home theatre PC initiatives which adds a whole lot more sense to the Windows 8 interface outside of the tablet PC space. This is a very positive outcome for the videogame industry as a whole since it offers the best of Xbox for those who desire it and the choice of the PC platform.
This however opposes Microsoft’s excessively strict stance on closed and proprietary video gaming platforms. Could Microsoft have been pushing their proprietary platform to gut the industry norms knowing that at some point they would roll back into their long-standing more-open nature of Windows? Could Microsoft be attempting to lock down PCs, meeting somewhere in the middle? We will see, but my hopes are that proprietary industry will finally move away from art. After all, why have a timeless classic if your platform will end-of-life in a half-dozen years at best?
Subject: Editorial, General Tech | July 8, 2011 - 12:29 AM | Scott Michaud
Tagged: mod, battlefield 3
The Battlefield franchise has had a somewhat indecisive history with the mod community. Battlefield 2 was developed in part by a mod team for the first game, Battlefield 1942, and mod tools were provided for several of their releases. Recently they shifted their focus on to the console spinoff, Bad Company. While the second in the franchise was created for the PC neither featured mod tools. Now that DICE has returned to the original canon with Battlefield 3 there were hopes that mod tools would return with the franchise but according to DICE that is not the case.
These tools are hard, just look at the destructibility, you wouldn’t like it…
German gaming site GameStar met up with DICE’s CEO Patrick Soderlund to discuss Battlefield 3. Soderlund answered an array of questions from the community about the Bad Company 2 friends list, alternatives to the commander mode, and the potential future of Mirror’s Edge. When questioned about the mod tools: Soderlund did not rule out the possibility of mod tools in the future but might as well done so. He contends that Frostbite 2 is too difficult to deal with for modders (which historically means: “the tools barely work for us, we are not going through the effort to polish them for public use”).
Surprisingly, to those who know me, I can agree with DICE’s stance on the issue. If your mod tools do not fit your level of polish required to release, then do not release them; provided, of course, you do not actively harm the creation of mods. With that in mind, the mod community is what will keep your game flowing with new content, for a little upfront cost. If your tail is shorter than you anticipated: this should be the first place to look.
Subject: Editorial, General Tech | July 7, 2011 - 04:06 AM | Scott Michaud
Tagged: starcraft 2, bumpday
This week the second season of Starcraft 2 came to a close immortalizing whatever rank you happened to be just prior to the lock. This is the second league freeze that Starcraft 2 underwent since it was launched last July. Starcraft 2 was a very anticipated launch for Blizzard due to the twelve year gap between it and its predecessor, Starcraft. Our forums were just one of many places where rabid fans frothed over Starcraft and I believe we deserve to nostalgia over it.
The calendar says 7/7/2011… … Bumpedaday??
Back in early 2001, just over ten years ago, ccrazy1263 created a very energetic forum thread questioning the coming of a new Starcraft. Blizzard would not even announce the existence of a new Starcraft until over 6 years later. One thing that comes with Starcraft is that players will constantly tell you their strategies, good or bad, and there were a few instances of that in this thread. In closing, the best advice to be had from this thread is short and sweet: Don’t play Hunter or Lost Temple against Hacker. They’ll kill your expansion.
Subject: Editorial, General Tech | July 5, 2011 - 02:06 PM | Scott Michaud
Tagged: windows, games
So I am a very avid supporter of PC gaming: I do not feel that I should pay license fees to a company to actively limit me. I feel that if people keep asking me whether HD-DVD or BluRay will win that there is no reason to have eight gaming platforms co-exist (more if you include Mac, iOS, Android, etc.). I feel that instead of relying on a cardboard coffin that will arrive in 1-2 business weeks, you or your friend should be allowed to fix your own hardware… or choose your own local small business computer store to deal with in person. I feel that it is much better to buy an extra GTX 560 every four years and have money left over for a meal than a multiplayer subscription pass that does not even let you run your own dedicated servers with admins (what cheaters? BAN!) So you can guess my reaction when I saw Microsoft roll Games for Windows Marketplace into Xbox.com.
Underlined for your convenience.
Now do not get me wrong, I was never much a fan of Games for Windows to begin with. Microsoft’s attempt to rid the PC of the stability stereotype was to push certification for all content on Games for Windows Live which worked very well for Fallout: New Vegas on the Xbox 360. Ironically the PC version was much more stable just after launch because the patch was stuck in certification on Xbox during the busy holiday season (lols!) The biggest problem with forcing certification is that would include mods as well (video, 3:10) and that is not something mod developers could really afford. Halo 2 Vista was one such Games for Windows Live game whose mod tools were so neutered that the provided tutorial level was impossible to create because jump pads assets were not provided nor able to be created.
Still, it seems odd to me for Microsoft to push so feverishly to roll PC gaming into Xbox branding when other initiatives like Steam are already showing the entire industry how to do things mostly right. It is possible that at some point Microsoft could roll Direct(X)box back in to Windows and simply create a canon home theatre PC (if Valve does not eat that lunch too); but if their plan is to merge Windows into Xbox then they are completely missing the point of why we would rather play a PC game with an Xbox 360 controller: because we choose to.
Subject: Editorial, General Tech | June 29, 2011 - 11:55 PM | Scott Michaud
Tagged: far cry, bumpday
This week Crytek released their DirectX 11 update to Crysis 2 and immediately hushed the crowing of all GPUs everywhere. We have gone through the various advancements in fairly great detail but that was not the first time our GPUs came Crying to us about the monster that goes Polybump in the night. No, our forums were aflame even before Crysis back when Far Cry was about to reach its Beta phase. I guess it is time to bump it up in our memory.
Back in early 2004 our forum-goers rallied around now defunct pre-beta screenshots of the best graphics they have ever seen… in 2004. People were surprised of the novelty of having mutant zombies in your game… in 2004. People were gathered to play with their fellow amdmb members… in 2004. People were gloating about how they cannot show what the beta looked like due to non-disclosure agreements… back in 2004. Of course we now know everything about Far Cry, its official sequel from another developer, and both of its spinoff sequels from its original developer. But obviously, what we know now we did not know then; let us nostalgia at the forum thread long since forgotten.
Subject: Editorial, General Tech | June 27, 2011 - 09:12 PM | Scott Michaud
Tagged: windows 8, leak
Update: 6/28/2011 - One of our commenters suggested that the screenshots were fake. Upon looking at ZDNet's sources -- it appears as if at least the first screenshot is fake (the tile screen) as well as their onscreen keyboard (which we did not link to). The other screenshots we linked follow a completely different aesthetic to the other screenshots on the fake portfolio (shape and color of close button, for instance) so they in fact appear to be genuine. Fooled me. -Scott
So Windows 8 was shown off at the All Things Digital D9 conference and surprise it was leaked. Naturally Microsoft did not show all aspects of the Windows 8 build at the conference; they must leave some cards hidden that are either not yet ready or otherwise not designed to be shown. Ziff Davis got a hold of someone who either had a leaked build of Windows 8 or otherwise access to screenshots that Microsoft did not intend to show. And what good are screenshots that are not in a slideshow?
Care to take a spin around the leek?
So we start off with the well-known start overlay with the typical tiles including weather, calendar, Computer, email, and Internet Explorer. The next image makes us feel immediately guilty for exactly a half of a second. The new interface extends all the way to the installer where you read the EULA and enter your personalization information. The windowing look and feel has changed with Windows 8 at least temporarily exaggerating the close button and minimizing the, well, minimize and full screen buttons. The ribbon UI is also seen exploding all across the interface including the file browser. Installations, at least of Windows software, are more integrated into the operating system. Lastly, the task manager is getting a facelift which may or may not be a bad thing.
What do you think of the leaked build? What would you do differently if you were Microsoft? (Registration not required to comment.)
Subject: Editorial, General Tech, Graphics Cards | June 27, 2011 - 04:44 PM | Scott Michaud
Tagged: dx11, crysis 2
Last Wednesday we reported on the announcement of the Crysis 2 DX11 patch and high resolution texture pack upcoming for the 27th of June. Looking at the calendar it appears as if your graphics card just ran out of time to rule the roost. Clocking in at 546 megabytes for the DirectX 11 update and 1695 megabytes for the high resolution texture pack the new updates are not small especially since that does not include the size of the 1.9 patch itself. The big question is whether these updates will push the limits of your computer, and if so, is it worth it?
Can you run me now? … Hello?
VR-Zone benchmarked the new updates on an Intel Core i7-965 system paired with an NVIDIA GeForce GTX 580. We believe they accidentally mislabeled their Extreme Quality benchmark with their Ultra Quality benchmark as the ultra is the more intensive of the two settings; also, ultra should have the biggest difference between DX9 and DX11 settings as DX11 effects are not enabled at the extreme settings. ((Update: 6/28/2011 - That's exactly what happened. VR-Zone fixed it; it is correct now.)) Under that assumption you are looking at approximately 40 FPS for a 1080p experience with that test system and all the eye-candy enabled. That is a drop of approximately 33% from its usual 60 FPS under extreme settings.
But how does it look? Read on for all of that detail.
Subject: Editorial, General Tech | June 25, 2011 - 02:09 PM | Scott Michaud
Tagged: mozilla, enterprise
For enterprise users looking to introduce Firefox to their business: you may wish to reconsider. Businesses are notorious for being substantially behind in version numbers, occasionally (or a lot) trading even security for compatibility. Mozilla had a staggered release schedule: a minor version number was little more than a security update; a major version number was a fairly-large overhaul. Enterprise users were able to upgrade minor version numbers and be reasonably assured that compatibility would be maintained. There were no such assurances for a major version number, thus requiring rigorous testing before applying. Mozilla has ended their policy of supporting back versions with security updates and are also moving between full versions much more rapidly, causing dissension amongst enterprise users.
Moving the world forward, not backwards, and always twirling towards freedom.
Ed Bott took the opportunity to prod Mozilla during his Thursday evening column. He contends that shutting out enterprise will assist in the impending implosion of Firefox and allow Microsoft and Google to pick up the pieces. I seriously disagree with that statement and applaud Mozilla for staying focused on their goal. True, Mozilla will be vastly less attractive to the enterprise; however, if Microsoft did not have Windows and Office to push with Internet Explorer, would search ad revenue and donations cover the long-term development cost incurred supporting enterprise users? And really, I would have thought Ed Bott of all people (ok, except maybe Paul Thurrott) would respect a company that can make a decision like Mozilla just did and stick by it after covering Microsoft for so long.
Subject: Editorial, General Tech | June 21, 2011 - 02:36 PM | Ryan Shrout
Tagged: VIA, sysmark, nvidia, Intel, benchmark, bapco, amd
It seems that all the tech community is talking about today is BAPCo and its benchmarking suite called Sysmark. A new version, 2012, was released just recently and yesterday we found out that AMD, NVIDIA and VIA have all dropped their support of the "Business Applications Performance Corporation". Obviously those companies have a beef with the benchmark as it is, yet somehow one company stands behind the test: Intel.
Everyone you know of is posting about it. My twitter feed "asplode" with comments like this:
AMD quits BAPCo, says SYSmark is nutso. Nvidia and VIA, they say, also. http://bit.ly/kHvKux
AMD: Voting For Openness: In order to get a better understanding of AMD's press release earlier concerning BAPCO... http://bit.ly/kNtKkj
Ooh, BapCo drama.
Even PC Perspective posted on this drama yesterday afternoon saying: "The disputes centered mostly over the release of SYSmark 2012. For years various members have been complaining about various aspects of the product which they allege Intel strikes down and ignores while designing each version. One major complaint is the lack of reporting on the computer’s GPU performance which is quickly becoming beyond relevant to an actual system’s overall performance. With NVIDIA, AMD, and VIA gone from the consortium, Intel is pretty much left alone in the company: now officially."
Obviously while cutting the grass this morning this is the topic swirling through my head; so thanks for that everyone. My question is this: does it really matter and how is this any different than it has been for YEARS? The cynical side of me says that AMD, NVIDIA and VIA all dropped out because each company's particular products aren't stacking up as well as Intel's when it comes to the total resulting score. Intel makes the world's fastest CPUs, I don't think anyone with a brain will dispute that, and as such on benchmarks that test the CPU, they are going to have the edge.
We recently reviewed the AMD Llano-based Sabine platform and in CPU-centric tests like SiSoft Sandra, TrueCrypt and 7zip the AMD APU is noticeably slower. But AMD isn't sending out press releases and posting blogs about how these benchmarks don't show the true performance of a system as the end user will see. And Intel isn't pondering why we used games like Far Cry 2 and Just Cause 2 to show the AMD APU dominating there. Why? Because these tests are part of a suite of benchmarks we use to show the overall performance of a system. They are tools which competent reviewers wield in order to explain to readers why certain hardware acts in a certain way in certain circumstances.
Continue reading for more on this topic...
Subject: Editorial, General Tech | June 20, 2011 - 03:24 AM | Tim Verry
Tagged: simulator, networking, Internet, cyber warfare
Our world is the host to numerous physical acts of aggression every day, and until a few years ago those acts have remained in the (relatively) easily comprehensible physical world. However, the millions of connected servers and clients that overlay the numerous nations around the world have rapidly become host to what is known as “cyber warfare,” which amounts to subversion and attacks against another people or nation through electronic means-- by attacking its people or its electronic and Internet-based infrastructure.
While physical acts of aggression are easier to examine (and gather evidence) and attribute to the responsible parties, attacks on the Internet are generally the exact opposite. Thanks to the anonymity of the Internet, it is much more difficult to determine the originator of the attack. Further, the ethical debate of whether physical actions in the form of military action is appropriate in response to online attacks comes into question.
It seems as though the Pentagon is seeking the answers to the issues of attack attribution and appropriate retaliation methods through the usage of an Internet simulator dubbed the National Cyber Range. According to Computer World, two designs for the simulator are being constructed by Lockheed Martin with a $30.8 million USD grant and Johns Hopkins University Applied Physics Laboratory with a $24.7 million USD grant provided by DARPA.
The National Cyber Range is to be designed to mimic human behavior in response to various DefCon and InfoCon (Informational Operations Condition) levels. It will allow the Pentagon and authorized parties to study the effectiveness of war plan execution as it simulates offensive and defensive actions on the scale of nation-backed levels of cyber warfare. Once the final National Cyber Range design has been chosen by DARPA from the two competing projects (by Johns Hopkins and Lockheed Martin), the government would be able to construct a toolkit that would allow them to easily transfer and conduct cyber warfare testing from any facility.
Image cortesy Kurtis Scaletta via Flickr Creative Commons.
Subject: Editorial, Graphics Cards, Processors, Mobile, Shows and Expos | June 16, 2011 - 02:41 PM | Ryan Shrout
Tagged: llano, liveblog, fusion, APU, amd, AFDS
The AMD Fusion Developer Summit 2011 is set to begin at 11:30am ET / 8:30am PT and promises to bring some interesting and forward looking news about the future of AMD's APU technology. We are going to cover the keynotes LIVE right here throughout the week so if you want to know what is happening AS IT HAPPENS, stick around!!
Subject: Editorial | June 15, 2011 - 09:54 PM | Scott Michaud
Tagged: duke nukem, bumpday
Yesterday saw the launch of Duke Nukem Forever. If you were a member of our forums for a long period of time you might actually have an odd case of deja vu. This is actually the second time that we have experienced the launch of Duke Nukem Forever. The first time was a sick joke one April fool’s day over nine years ago which is quite fitting for the lewd and crude franchise. Some argue that our current release is a sick joke, but that is a whole other story.
Duke Nukem: Forever and 3363 days.
Back in April 1st, 2002: Think Geek announced the availability of Duke Nukem Forever for Linux, Windows 3.1, WheatoniX, and Plan 9. The general buzz on the forum was about the marvel of being a PC gamer and how much of an impact Duke Nukem 3D had on our lives. So go on, nostalgia at the old fashioned giant PC gaming boxes and nostalgia at the forum thread long since forgotten.
Subject: Editorial, General Tech, Shows and Expos | June 15, 2011 - 05:58 PM | Ryan Shrout
Tagged: programming, microsoft, fusion, c++, amp, AFDS
During this morning's keynote at the AMD Fusion Developer Summit, Microsoft's Herb Sutter went on stage to discuss the problems and solutions involved around programming and developing for multi-processing systems and heterogeneous computing systems in particular. While the problems are definitely something we have discussed before at PC Perspective, the new solution that was showcased was significant.
C++ AMP (accelerated massive parallelism) was announced as a new extension to Visual Studio and the C++ programming language to help developers take advantage of the highly parallel and heterogeneous computing environments of today and the future. The new programming model uses C++ syntax and will be available in the next version of Visual Studio with "bits of it coming later this year." Sorry, no hard release date was given when probed.
Perhaps just as significant is the fact that Microsoft announced the C++ AMP standard would be an open specification and they are going to allow other compilers to integrated support for it. Unlike C# then, C++ AMP has a chance to be a new dominant standard in the programming world as the need for parallel computing expands. While OpenCL was the only option for developers that promised to allow easy utilization of ALL computing power in a computing device, C++ AMP gives users another option with the full weight of Microsoft behind it.
To demonstrate the capability of C++ AMP Microsoft showed a rigid body simulation program that ran on multiple computers and devices from a single executable file and was able to scale in performance from 3 GLOPS on the x86 cores of Llano to 650 GFLOPS on the combined APU power and to 830 GFLOPS with a pair of discrete Radeon HD 5800 GPUs. The same executable file was run on an AMD E-series APU powered tablet and ran at 16 GFLOPS with 16,000 particles. This is the promise of heterogeneous programming languages and is the gateway necessary for consumers and business to truly take advantage of the processors that AMD (and other companies) are building today.
If you want programs other than video transcoding apps to really push the promise of heterogeneous computing, then the announcement of C++ AMP is very, very big news.
Subject: Editorial, Processors, Shows and Expos | June 14, 2011 - 05:09 PM | Ryan Shrout
Tagged: nvidia, Intel, heterogeneous, fusion, arm, AFDS
Before the AMD Fusion Developer Summit started this week in Bellevue, WA the most controversial speaker on the agenda was Jem Davies, the VP of Technology at ARM. Why would AMD and ARM get together on a stage with dozens of media and hundreds of developers in attendance? There is no partnership between them in terms of hardware or software but would there be some kind of major announcement made about the two company's future together?
In that regard, the keynote was a bit of a letdown and if you thought there was going to be a merger between them or a new AMD APU being announced with an ARM processor in it, you left a bit disappointed. Instead we got a bit of background on ARM how the race of processing architectures has slowly dwindled to just x86 and ARM as well as a few jibes at the competition NOT named AMD.
As is usually the case, Davies described the state of processor technology with an emphasis on power efficiency and the importance of designing with that future in mind. One of the interesting points was shown in regard to the "bitter reality" of core-type performance and the projected DECREASE we will see from 2012 onward due to leakage concerns as we progress to 10nm and even 7nm technologies.
The idea of dark silicon "refers to the huge swaths of silicon transistors on future chips that will be underused because there is not enough power to utilize all the transistors at the same time" according to this article over at physorg.com. As the process technology gets smaller then the areas of dark silicon increase until the area of the die that can be utilized at any one time might hit as low as 10% in 2020. Because of this, the need to design chips with many task-specific heterogeneous portions is crucial and both AMD and ARM on that track.
Those companies not on that path today, NVIDIA specifically and Intel as well, were addressed on the below slide when discussing GPU computing. Davies pointed out that if a company has a financial interest in the immediate success of only CPU or GPU then benchmarks will be built and shown in a way to make it appear that THAT portion is the most important. We have seen this from both NVIDIA and Intel in the past couple of years while AMD has consistently stated they are going to be using the best processor for the job.
Amdahl's Law is used in parallel computing to predict the theoretical maximum speed up using multiple processors. Davies reiterated what we have been told for some time that if only 50% of your application can actually BE parallelized, then no matter how many processing cores you throw at it, it will only ever be 50% faster. The heterogeneous computing products of today and the future can address both the parallel computing and serial computing tasks with improvements in performance and efficiency and should result in better computing in the long run.
So while we didn't get the major announcement from ARM and AMD that we might have been expecting, the fact that ARM would come up and share a stage with AMD reiterates the message of the Fusion Developer Summit quite clearly: a combined and balanced approach to processing might not be the sexiest but it is very much the correct one for consumers.
Subject: Editorial, General Tech | June 9, 2011 - 12:25 AM | Scott Michaud
Tagged: windows 8, silverlight
That interface doesn’t look very silvery, or light.
I think the real message here is that when you invest (through time, money, or otherwise) in a proprietary infrastructure you need to expect that you have no real recourse should the owner work against you; you voided all recourse except for what is explicitly contractually bound to you. In the case of an open, particularly copyleft, platform: should support from the original owners be absent or insufficient you are legally allowed to take over provided that right is also granted by you. Often it may still be worthwhile to invest in proprietary platforms, but remember, you give up your right to maintain your dependencies. All your dependent art is relying on your trust in the platform owner, and you have no legal recourse, because you gave it away.
Do you have any comments on this? Discuss below.
Subject: Editorial, General Tech, Shows and Expos | May 29, 2011 - 06:08 PM | Ryan Shrout
Tagged: pcper, computex 2011, computex
It is that time of year again where the staff at PC Perspective wades through the moist air of Taipei, Taiwan to bring you the latest news, videos and updates from the world of technology and computing hardware. This marks my 10th appearance on the island of Taiwan, but Ken's first so it should be an interesting experience for both of us.
Yes, this scary girl is here again too.
This year we have a surprising amount of expected hardware to investigate starting with an onslaught of AMD-based technology. The AMD 990FX chipset will be out and about in force from the likes of ASUS, MSI, ECS and others while the Fusion APUs will be flaunted in notebooks and likely even some "surprise" Llano designs. Intel's Z68 chipset, though already launched, will still be a focus but the company will surprise many with a dramatic move into the world of mobile computing with the Atom line.
Temperatures are high but so is the excitement!!
Subject: Editorial, General Tech | May 27, 2011 - 06:41 PM | Jeremy Hellstrom
Tagged: friday, forum
Our Tech Talk Forum is a great place to hang out and discuss general technology questions, such as the eternal question of which screwdriver to use, since we can't get hold of a working Sonic Screwdriver, do you have a preferred electric screwdriver or do you prefer the traditional one like mpulliam and I do? Perhaps you are more interested in moving to SoCal where NewEgg is apparently opening up a pick up location.
Head over to the Processor Forum for speculation on Bulldozer, since the recent leaks have given us a bit more information; or maybe a laugh at the Apple fanatics and the price they pay for the same hardware we use? Especially considering that while some of their designs are nice, they can never match the imagination of a dedicated modder!
There is a lot more inside, take a walk through the forums and see what you can find or even better ... what you can offer! For a more laid back way to catch up on the latest tech news you should catch the new PC Perspective Podcast, #156 even has Ryan in it!