All | Editorial | General Tech | Graphics Cards | Networking | Motherboards | Cases and Cooling | Processors | Chipsets | Memory | Displays | Systems | Storage | Mobile | Shows and Expos
Subject: General Tech | June 16, 2011 - 12:57 PM | Jeremy Hellstrom
Tagged: amd, Intel, nvidia
In some sort of bizarre voyeuristic hardware love/hate triangle AMD, Intel and NVIDIA are all semi-intertwined and being observed by Microsoft. Speaking with The Inquirer the VP of product and platform marketing at AMD, Leslie Sobon, stated that there was no chance that Intel would attempt to purchase NVIDIA as AMD did with ATI. AMD's purchase was less about the rights to the Radeon series as it was taking possession of the intellectual property that ATI owned after a decade of creating GPUs and lead directly to the APUs that AMD has recently released which will likely become their main product. Intel already has a working architecture that combines GPU and CPU and doesn't need to purchase another company's IP in order to develop that type of product.
There is another reason for purchasing NVIDIA though, which has very little to do with their discreet graphics card IP and everything to do with Tegra and Fermi which are two specialized products which so far Intel doesn't have an answer for. A vastly improved and shrunken Atom might be able to push Tegra off of mobile platforms and perhaps specialized SandyBridge CPUs could accelerate computation like the Fermi products do but so far there are no solid leads, only speculation.
If you learn more from your failures than your successes then Intel knows a lot about graphics.
"CHIP DESIGNER AMD believes that it is on a divergent path from Intel thanks to its accelerated processor unit (APU) and that Intel buying Nvidia "would never happen"."
Here is some more Tech News from around the web:
- Find Out if Your Passwords Were Leaked by LulzSec Right Here @ Gizmodo
- Adobe patches critical bugs in Flash and Reader @ The Register
- Umi, we hardly knew ye: contemplating the fate of the videophone in 2011 @ Ars Technica
- 'A SHARK attacked my ROBOT', gasps ex-Sun exec @ The Register
- We’ve got a real bone to pick with this mouse @ Hack a Day
- Fun Quotes from the AFDS Media Roundtable @ SemiAccurate
Subject: General Tech, Shows and Expos | June 15, 2011 - 09:14 PM | Scott Michaud
Tagged: opencl, amd, AFDS
If you are a developer of applications which requires more performance than a CPU alone can provide then you are probably having a gleeful week. Today Microsoft announced their competitor to OpenCL and we have a large write-up about that aspect of their keynote address. If you are currently an OpenCL developer you are not left out, however, as AMD has announced new tools designed to make your life easier too.
General Purpose GPU utilities: Because BINK won't satisfy this crowd.
(Logo trademark Apple Inc.)
AMD’s spectrum of enhanced tools includes:
- gDEBuger: An OpenCL and OpenGL debugger, profiler, and memory analyzer released as a plugin for Visual Studio.
- Parallel Path Analyzer (PPA): A tool designed to profile data transfers and kernel execution across your system.
- Global Memory for Accelerators (GMAC) API: Lets developers use multiple devices without needing to manage multiple data buffers in both the CPU and the GPU.
- Task Manager API: A framework to manage scheduling kernels across devices.
These tools and utilities should make the development of software easier and allow more developers to take the risk on the new technology. The GPU has already proven itself worthy of more and more important tasks and it is only a matter of time before it is finally ubiquitous enough that it is a default component as important as the CPU itself. As an ironic aside, that should spur the adoption of PC Gaming given how many people would have sufficient hardware.
Subject: Editorial, General Tech, Shows and Expos | June 15, 2011 - 05:58 PM | Ryan Shrout
Tagged: programming, microsoft, fusion, c++, amp, AFDS
During this morning's keynote at the AMD Fusion Developer Summit, Microsoft's Herb Sutter went on stage to discuss the problems and solutions involved around programming and developing for multi-processing systems and heterogeneous computing systems in particular. While the problems are definitely something we have discussed before at PC Perspective, the new solution that was showcased was significant.
C++ AMP (accelerated massive parallelism) was announced as a new extension to Visual Studio and the C++ programming language to help developers take advantage of the highly parallel and heterogeneous computing environments of today and the future. The new programming model uses C++ syntax and will be available in the next version of Visual Studio with "bits of it coming later this year." Sorry, no hard release date was given when probed.
Perhaps just as significant is the fact that Microsoft announced the C++ AMP standard would be an open specification and they are going to allow other compilers to integrated support for it. Unlike C# then, C++ AMP has a chance to be a new dominant standard in the programming world as the need for parallel computing expands. While OpenCL was the only option for developers that promised to allow easy utilization of ALL computing power in a computing device, C++ AMP gives users another option with the full weight of Microsoft behind it.
To demonstrate the capability of C++ AMP Microsoft showed a rigid body simulation program that ran on multiple computers and devices from a single executable file and was able to scale in performance from 3 GLOPS on the x86 cores of Llano to 650 GFLOPS on the combined APU power and to 830 GFLOPS with a pair of discrete Radeon HD 5800 GPUs. The same executable file was run on an AMD E-series APU powered tablet and ran at 16 GFLOPS with 16,000 particles. This is the promise of heterogeneous programming languages and is the gateway necessary for consumers and business to truly take advantage of the processors that AMD (and other companies) are building today.
If you want programs other than video transcoding apps to really push the promise of heterogeneous computing, then the announcement of C++ AMP is very, very big news.
If you happened to open up the store page in the Steam client or glance at their website, you may have noticed that Steam has made a moderately big announcement. Valve's digital download service now supports Free-to-Play games, which are games that are free to download and play at the basic level; however aesthetic and other upgrades can be purchased via so-called "microtransactions". F2P games on still will be free to download and will not require a credit card to do so.
Steam seems excited about the new F2P games.
At launch, the service is featuring five new Free-to-Play games including Champions Online: Free For All, Spiral Knights, Global Agenda: Free Agent, Forsaken World, and Alliance of Valiant Arms. According to the F2P Steam FAQ, games in which you wish to purchase content will be done through the use of your Steam Wallet. Further, for any Steam account that does not have at least one purchased (non Free-to-Play) game or a funded Steam Wallet will be considered a "Limited User" and will be restricted in the community features that it is able to access. Specifically, limited users can create community groups, be added as friends, and chat with other users; however, they are not able to send out friend invitations or start chat sessions (a non-limited user must initiate chat).
In adding the new genre to its repertoire, Steam will greatly increase its digital games library and add more options for PC gamers. One game that I have not played in some time that I would love to see make its way onto the new Free-to-Play Steam selection is a FPS game called Crossfire. That game was a good example of Free-To-Play done right as even accounts that did not spend a dime where able to stay competitive. Is there a Free-to-Play game that you would like to see Steam feature, and do you think F2P will add value to the service? Let us know in the comments.
Subject: General Tech | June 15, 2011 - 12:46 PM | Jeremy Hellstrom
Tagged: gaming, duke nukem, 10 commandments
We need a new joke, the poster boy of vapourware has actually arrived and no one remembers the Phantom console. You can catch up on all of the reviews of Duke Nukem Forever below the fold, but make sure you don't say anything mean about the game or the PR firm will get you. There is also a lot of previews from E3 to drool over, many new games offered teases of their unreleased products.
Before you take a look at the games, The Tech Report has recently crafted 10 commandments that all PC games should follow. Read through them and see which of the new games look to be following the reasonable requirements that they have listed.
It's beside the Any key, right?
"Picture this for a second: you just unpacked the latest PlayBox 720-X blockbuster game, Gran Gears of Duty Fantasy XVIII. It's a game so juicy and dreamy that it'll send you flying into all the colors of the rainbow, twitching and jerking with pleasure-induced spasms just from looking at the loading screen. Let's assume for the sake of argument that said game is a first-person shooter, like, oh, about 135% of recent releases. You insert the Megaray disc, go about the installation process, and merrily start to play.
All of a sudden, you notice the left stick is used for switching weapons. The right stick moves the character, and shooting is only accomplished by pressing it. The camera is moved with the directional buttons, and the triangle, square, A, and B buttons are used for your character's smartass quips. You enter the menu to change the controls, but you can only navigate them using the motion sensors. After five minutes of furniture-dusting motions, you finally enter the options menu and find out there are barely any options, and none that matter. Frustrated, you throw the TenAxis controller at your 4D TV screen and take the shiny disc out of the console to find out whether it will blend."
Here is some more Tech News from around the web:
- Duke Nukem Forever Performance Test @ TechSpot
- Wot I Think: Duke Nukem Forever @ Rock, Paper, SHOTGUN
- Duke Nukem Forever Game Review (PC) @ HardwareHeaven
- Duke Nukem Forever: barely playable, not funny, rampantly offensive @ Ars Technica
- Duke Nukem Forever Review @ Techgage
- Skyrim's dragon battles top off impressive demo @ Ars Technica
- Asura's Wrath is effortlessly weird and weirdly awesome @ Ars Technica
- Gaming Friday – Bulletstorm @ ThinkComputers
- Serious Sam 3 adds no modern gameplay, thank heavens @ Ars Technica
- Star Wars: The Old Republic's lead writer on good Sith, evil Jedi @ Ars Technica
- Section 8: Prejudice Review (PC) @ HardwareHeaven
- DiRT 3 Gameplay Performance @ The Register
- Assassin's Creed Revelations: The Escape From Constantinople Q&A @ HEXUS
- CANVAS Teaser Is Sinister, Awesome @ Rock, Paper, SHOTGUN
- Wot I Think: Red Faction Armageddon @ Rock, Paper, SHOTGUN
- Steam Now Offering Free-To-Play Games @ Slashdot
- A character, not a voice: tuning the narration in Bastion @ Ars Technica
- Silent Hill: Downpour - Xbox 360, PS3 @ HEXUS
- Wii? Maybe U... But I'm Not Sold Yet @ Techgage
- Super Mario 3DS hands-on: the Tanooki suit is back! @ Ars Technica
- Dead or Alive Dimensions 3DS @ Tweaktown
- Duke Nukem Forever Xbox 360 @ Tweaktown
- Icebreaker Hockey for iOS: hockey stripped down to $1 perfection @ Ars Technica
Subject: General Tech | June 15, 2011 - 12:16 PM | Jeremy Hellstrom
Tagged: servers, calxeda, arm
ARM has assembled their own Super Best Friends in a team lead by Calxeda, and composed of Autonomic Resources, Canonical, Caringo, Couchbase, Datastax, Eucalyptus Systems, Gluster, Momentum SI, Opscode, and Pervasive. This places Ubuntu as the ARM OS of choice for the server room and as it includes companies developing applications for running Cloud services, not only Microsoft should be paying attention; applications like Amazon's EC2 could face new competition as well.
Calexda's current reference machines pack 120 server nodes with 480 cores in a 2U chassis, a density which even a 1W Atom is going to find hard to match and the 1W Atoms are still a ways away. They are planning on getting the machines out to clients for testing by the end of the year, Intel's time table is nowhere near that tight. Read more about the low powered battle for dominance at The Register.
"With Intel's top brass bad-mouthing ARM-based servers, upstart server chip maker Calxeda can't let Intel do all the talking. It has to put together an ecosystem of hardware and software partners who believe there's a place for a low-power, 32-bit ARM-based server platform in the data center."
Here is some more Tech News from around the web:
- ARM Fellow takes the mic at AMD event @ The Tech Report
- AMD demos Trinity laptop @ SemiAccurate
- Epson WorkForce 60 Review @ TechReviewSource
- Chameleon WiGig gains momentum @ The Register
- Having fun with Microsoft Windows @ t-break
- Microsoft squeaks on Google Nortel sale @ The Register
- A tour of Zotac's Dongguan factory @ The Tech Report
Subject: General Tech | June 14, 2011 - 05:29 PM | Jeremy Hellstrom
Tagged: audio, razer, ferox, 2.1
The Razer Ferox speakers are designed to be portable, a pair of satellites measuring 70x70x64mm (not even 3") which come in a handy carrying case. They sport batteries that should last about 11 hours that are recharged over a USB connection but still require a 3.5mm jack to carry the audio, something that did not impress t-break in the least. The sound quality was good for this type of speaker, which equates to unnoticeable bass and decent mid and high end when in use. If you usually use headphones and simply need a way to share your audio, as opposed to needing new speakers then check out the Ferox, otherwise Razer has better choices as do Corsair and other manufacturers.
"Razer is no stranger to high quality audio equipment, what with the number of high-end stereo and surround headsets over the past years. Their breakthrough hit, the Razer Mako 2.1 THX speakers were one of the best desktop audio speakers at the time, and are still hard pressed to beat till this day. And now with the new Ferox speakers, Razer has entered the world of mobile speakers with a big bang."
Here is some more Tech News from around the web:
- Radiopaq Duo Headphones Review @ Tech-Reviews
- Thermaltake Shock Headset @ Bjorn3D
- Apacer Audio Steno AU825 MP4 Player Review @ Real World Labs
- SteelSeries 7H Headset for iPod, iPhone and iPad Review @ HardwareHeaven
- Speedlink Xbox 360 Headset Adapter @ XSReviews
Subject: General Tech, Systems | June 14, 2011 - 03:24 PM | Scott Michaud
Tagged: llano, hp
Level up! Llano life increased by 11 HP.
So, AMD is currently having a little shindig right now as you might be aware from recent news posts and news is just a leaking from the rafters. HP recently contacted us to announce that they just expanded both their consumer and business product lines to include 11 new models using “AMD’s latest Vision Technology”. What this means is we can expect a large array of products coming from HP that utilizes the latest generation of AMD CPUs and GPUs from their new Llano-based AMD A-Series product line. Expect a helping of Llano on your HP in the near future.
Subject: General Tech | June 14, 2011 - 12:02 PM | Jeremy Hellstrom
Tagged: http, tcp, spdy, Internet
Google has been working on SPDY, a new protocol which is intended to speed up HTTP without forcing changes to existing websites or protocols. This application-layer protocol sits between HTTP and TCP, replacing neither instead translating for the application layer and the transport layer to optimize certain parts of the transaction. Specifically they hope to allow multiple connections over TCP, something that up until now is provided by a workaround in the browser which creates parallel connections as well as getting servers to push data to clients more effectively. They are also working to reduce latency by reducing the size of the headers that are transported which will be very important in the near future, not only as a way to speed up SSL connections but to help with the increased size of IPv6 headers.
Up until now SPDY has only been available for Chrome and even then only for certain Google sites which utilize the new translation protocol. Now Strangeloop is offering an online service as well as hardware which will allow you to implement SPDY without the need to change your website or host. The Register covers the long overdue change to TCP here.
"Strangeloop – a Vancouver-based outfit offering an online service for accelerating website load times – has embraced Google's SPDY project, a new application-layer protocol designed to significantly improve the speed of good ol' HTTP."
Here is some more Tech News from around the web:
- Intel: Chinese microprocessor development inefficient @ SemiAccurate
- Intel new server platform expected to start large replacement trend @ DigiTimes
- Games co Epic resets passwords after hack attack @ The Register
- Research @ Intel: The cloud's future is many-core and GPU accelerated @ Ars Technica
- Planar structure extends lifetime of memristor @ Nanotechweb
- Nikon COOLPIX S570 12MP Digital Camera Review @ ThinkComputers
Subject: General Tech, Processors | June 14, 2011 - 02:47 AM | Scott Michaud
Tagged: Intel, haswell
Intel’s new processor lines come in two flavors: process shrinks and new architectures. Each revision comes out approximately a year after the prior one alternative between new architectures (tock) and process shrinks (tick). Sandy Bridge was the most recent new architecture which will be followed by Ivy Bridge, a process shrink of Sandy Bridge, and that will be succeeded by Intel’s newest architecture: Haswell.
I can Haswell?
The instructions added by Intel for their upcoming Haswell architecture are useful for a whole range of applications from image and video processing; to face detection; to database manipulation; to the generation of hashes; as well as arithmetic in general. As you can see the addition of instructions in this revision is quite wide in its scope. Keep in mind that the introduction of a new instruction set does not mean that programs will be optimized to take advantage of the added benefits for some time. However, when programs do start optimizing for the newer architectures it looks as though Haswell’s new offerings will speed up otherwise complicated tasks into a single instruction.
What task would you like to see a speedup on? Comment below.
Subject: General Tech | June 13, 2011 - 01:52 PM | Jeremy Hellstrom
Tagged: mechanical keyboard, input, steelseries
Mechanical keyboards seem to be a hot topic, with round ups appearing to deal with all of the new boards coming out. Hardware Heaven chose to focus on one particular product, the SteelSeries 6Gv2 Mechanical gaming keyboard, which thankfully didn't take 'gaming' to mean sticking extra buttons all along the side. The Cherry Black MX designed keys are very common amongst these new mechanical keyboards though the n-key rollover, being able to hit an unlimited number of keys and have them properly register, is not something you find on all USB keyboards. The 6Gv2 can handle multiple keys for you circle strafers and replacing the Windows key on the left hand side with a 'media key' that is disabled in games is a very nice touch. Check out the full review at Hardware Heaven since there are some negative aspects to the design of this board.
"For quite some time the gaming keyboard market has concentrated on products which add macro buttons, re-assignments, profiles, USB and audio pass-through and weighted key actions to enhance the gaming experience. In addition to this we see branded products such as the Razer StarCraft 2 gear and SteelSeries Medal of Honor products however few manufacturers have looked to release high quality mechanical keyboards for the gaming masses.
There have been a few though and these have clearly made an impact with gamers as we are regularly seeing manufacturers launch their own mechanical gaming models. One manufacturer which has historically offered mechanical keyboards for gamers is SteelSeries and they are now back with a new model, the 6Gv2 which we have connected to our system today."
Here is some more Tech News from around the web:
- Steelseries Ikari Laser Mouse (White) Review @ t-break
- Cooler Master CM Storm Spawn Gaming Mouse Review @ Tweaknews
- SteelSeries Cataclysm MMO Gaming Mouse Review @ Real World Labs
- Zowie G-TF Speed Review @ XSReviews
Subject: General Tech | June 13, 2011 - 11:47 AM | Jeremy Hellstrom
Tagged: microsoft, patch tuesday, security, windows, internet explorer, silverlight
Tomorrow will see the arrival of 9 critical security patches and 7 recommended ones, covering Windows, IE, Silverlight and Office. The critical patches all resolve remote code execution vulnerabilities, the recommended vary from the same type as well as privledge escalation and denial of service vulnerabilities. WinXP through Win7 as well as server OSes will all be affected so be warned that your Tuesday and Wednesday might not be very fun. Follow the link from The Register to see Microsoft's pre-release document for yourself.
Adobe, obviously not wanting to seem lazy, is also pushing out a patch for both Reader and Acrobat.
"Microsoft is preparing a bumper Patch Tuesday for next week, with 16 security bulletins that collectively address 34 vulnerabilities.
Nine of the bulletins earn the dread rating of critical, while the other seven grapple with flaws rated as important. All supported versions of Windows will need patching on 14 June along with various server-side software packages and applications, including the .NET framework and SQL Server. Internet Explorer, which is affected by two bulletins, will also need some fiddling under the bonnet."
Here is some more Tech News from around the web:
- Why Microsoft has made developers horrified about coding for Windows 8 @ Ars Technica
- PC Mark 7 Performance Review @ OCC
- PathScale Open-Sources The EKOPath 4 Compiler Suite @ Phoronix
- Samsung Galaxy Tab sneak peak at Dubai @ t-break
- HIS Solar LED Flashlight @ Benchmark Reviews
- Sumo Lounge Titan @ Phoronix
- Final Benchmarks Of Project Dirndl @ Phoronix
- Tablet Wars Single Stage Phase and Computex @ NinjaLane
- Win a HIS HD 5670 IceQ 1GB Graphics Card @ eTeknix
Subject: General Tech | June 13, 2011 - 04:15 AM | Scott Michaud
Tagged: mechanical keyboard, cherry
There is a large amount of choice when it comes to PC components and input devices are no exception to that assertion. You are probably well aware of the multitude of choices when it comes to non-standard mice in terms of number of buttons and resolution of the optical and/or laser sensor. Keyboards have their own higher performance counterparts as well: not just in terms of how many web and media function buttons can be crammed on them, but also how the keys themselves register a press. Recently Tom’s Hardware reviewed a series of mechanical keyboards based on their switches and gave a lot of background information about what advantages and disadvantages each switch has.
Are you a mechanical keyboard virgin? Feeling the MX Blues?
(Logo from the Cherry Corporation)
My first couple keyboards were the old IBM model M buckle spring keyboards. Eventually when I got a later computer I moved on to the cheap keyboards and immediately missed my original mechanical keyboards. Years and a little shopping around later, I eventually settled on the Logitech G15v1 as my first attempt at a higher-end gaming keyboard. It was with the G15v1 that I experienced serious limitations to be had with some, particularly non-mechanical, keyboards: I am a left-handed gamer. The Logitech G15v1 was optimized for right handed gamers as a lot of arrow-key combinations with shift or control did not register by the keyboard; Logitech expected, when they designed the keyboard, that everyone’s mouse would be on the right of the keyboard, and thus the further away WSAD keys would be used. Consider playing as a Scout in Team Fortress 2 but not being able to jump sideways and only being able to crouch-walk in a straight line. While each keyboard is designed with a different set of jammable key combinations it was events like those that led me to go overkill and purchase a mechanical keyboard with NKRO attached via PS/2 port.
Do you have any keyboard stories? Comment below. Otherwise, check out Tom’s Hardware’s guide and review to mechanical keyboards.
Subject: General Tech, Displays | June 12, 2011 - 05:56 PM | Scott Michaud
Tagged: SMART, 3d
SMART has been making interactive whiteboards for quite some time now. An interactive whiteboard is essentially a giant writing tablet similar to a Wacom. This tablet is also a projector screen which is often wall mounted but could be mounted on a cart. SMART Boards attach to PCs by USB and could attach to video and audio out if you purchase one with an attached projector and speakers rather than use your own. Recently SMART announced and released their fifth generation product line complete with a projector supporting HDMI input and active 3D technology.
IT’S LIKE I CAN TOUCH YOU!
(Image by SMART Technologies)
While I can see this useful for companies that are doing 3D technology during their company, investor, and vendor meetings it seems a little bit unlikely that active 3D will appear in the classroom. It seems quite difficult for me to imagine twenty to forty students each with their own active shutter 3D glasses atop the investment of the 3D interactive whiteboard itself. Also while it might be to support the 3D functionality of the projector it seems quite odd to include HDMI functionality and barely exceed 720p resolution (1280x800) in your highest-end projector.
Subject: General Tech | June 12, 2011 - 01:12 PM | Tim Verry
Tagged: US, technology, networking, IT
The US has seen a rather rapid rise in unemployment in the last few years as companies cut back on staff and computing costs. According to Computer World, Tom Silver has been quoted in saying “several years ago companies cut back pretty far, particularly in infrastructure and technology development.” Silver further believes that the tech unemployment rate is half that of the national unemployment rate due to companies needing to replace aging hardware, software, and deal with increased security threats. 65% of 900 respondents in a recent biannual hiring survey conducted by Dice.com found that hiring managers and head hunters plan on bringing even more new workers into their businesses in the second half of 2011 versus the first half.
Workers with mobile operating system, hardware, and ecosystem expertise and java development skills are the most desirable technology workers, according to Computer World. Although anyone with an IT background and recent programming skills have a fairly good chance of acquiring jobs in a market that is demanding now-rare talent. Employers are starting to be more confident in the economy and thus are more willing to invest in new workers. In an era where Internet security is more important that ever, skilled enterprise IT workers are becoming a valuable asset to employers, who are increasingly fighting for rare talent and incentivizing new workers with increased salaries.
Even though businesses are still remaining cautious in their new hiring endeavors, it is definitely a good sign for people with tech backgrounds who are looking for work as the market is ever so slowly starting to bounce back. For further information on the study, Computer World has the full scoop here.
Are you in or studying to enter into the IT profession? Do you feel confident in the US employers' valuation of their IT workers?
Subject: General Tech | June 12, 2011 - 10:36 AM | Tim Verry
Tagged: networking, dell, cloud computing
A recent survey conducted during the first two days of the Cloud Expo by Marketing Solutions and sponsored by Dell suggests that IT professionals believe that their less technical CEOs believe cloud computing to be a "fad" that will soon pass. On the other hand, IT departments see the opportunities and potential of the technology. This gap between the two professions, according to Dell, lies in "the tendency of some enthusiasts to overhype the cloud and its capacity for radical change." Especially with a complex and still evolving technology like cloud computing, CEOs are less likely to see the potential benefits and moreso the obstacles and cost to adopt the methods.
The study surveyed 223 respondents from various industries (excluding technology providers), and found that the attitudes of IT professionals and what they felt their respective CEOs' attitudes were regarding "the cloud" were rather different. The pie graphs in figure 1 below illustrate the gap between the two professions mentioned earlier. Where 47% of those in IT see cloud computing as a natural evolution of the trend towards remote networks and virtualization, only 26% of IT believed that CEOs agreed. Also, while 37% of IT professions stated that cloud computing is a new way to think about their function in IT, "37 percent deemed their business leaders mostly likely to describe the cloud as having “immense potential,” contrasted with only 22 percent of the IT pros who said that was their own top descriptor."
Further, the survey examined what both IT professionals and CEOs believed to be obstacles in the way of adopting cloud computing. On the IT professionals' front, 57% believed data security to be the biggest issue, 32% stated industry compliance and governance as the largest obstacle, and 27% thought disaster recovery options to be the most important barrier, contrasted with 51%, 30%, and 22% of CEOs. This comparison can be seen in figure 2 below.
While the survey has handily indicated that enterprises' IT departments are the most comfortable with the idea of adopting cloud computing, other areas of the business could greatly benefit from the technology but are much more opposed to the technology. As seen in figure 3, 66% of IT departments are willing to advocate for cloud computing, only 13% of Research and Development, 13% of Strategy and Business Development, and a mere 5% of Supply Chain Management departments feel that they would move to cloud computing and benefit from the technology.
Dell stated that IT may be able to help in many more functions and departments by advocating for and implementing cloud computing strategies in information-gathering and data-analyzation departments. In doing so, IT could likely benefit the entire company and further educate their CEOs in cloud computing's usefulness to close the gap between the IT professionals' and CEO's beliefs.
You can read more about the Dell study here. How do you feel about cloud computing?
Subject: General Tech | June 12, 2011 - 04:08 AM | Scott Michaud
Tagged: windows 8, ImmersiveUI
Microsoft announced and demonstrated their Windows 8 interface a couple of weeks ago and since then there has been some love and some hate for it by various groups. The idea that the new paradigm for icons would display information from the program, particularly in such a fashion, better suits a tablet rather than a traditional desktop interface. Regardless, there would likely be some application for such an interface and you do not need Windows 8 to unofficially have it.
“Start”: must be Windows.
ImmersiveUI developer Sergio James Bruccoleri has released a video to show his pre-beta interface for Windows 7. In his demonstration he showed various websites and programs launched with a little bit of feedback in the tiles such as his Facebook name and Xbox Live gamertag with avatar. Bruccoleri has stated that a public beta is forthcoming with “effects and some cool stuffs.”
Would you find yourself adding this to your Windows desktop? If so, on what device?
Subject: General Tech | June 11, 2011 - 11:21 PM | Tim Verry
Tagged: wintel, microsoft, Intel, asustek
DigitTimes reports that the so called “Wintel” era is over. With Wintel representing the fusing of a Windows operating system on Intel x86 processors, Asustek Jonney Shih believes that the time period where Windows and Intel processors dominated the PC, tablet PC, and handset markets have passed. This is due in part to the rise of Android and ARM on the mobile front and increased mind share (and in some cases competitive market share) of the Mac OSX and iOS ecosystems on the PC and mobile platforms respectively. Shih further stated that the rising market share of once-smaller operating systems from competitors encourages healthy competition and innovation in the industry.
As mobile hardware advances to once-unprecedented levels of performance, Asustek sees the lines between what constitutes mobile handsets, ultra-portable computing devices and traditional computers breaking down. All these devices will soon start to coalesce into a new IT market where computing is more about productivity and entertainment more so than choosing differing classes of hardware as they will all be “good enough” machines.
DigiTimes states that the rise of the tablet PC will likely increase manufacturers abilities to try new things and sell numerous units; however, it will also impact and “significantly reshuffle the ranking of the whole IT market.”
With Microsoft currently commanding approximately 88.69% of the client OS market share (according to Net Market Share at time of writing), and Intel being the leading manufacturer of x86 CPUs, the “Wintel” relationship still has a good deal of weight to throw around and influence the market; however, on the mobile front the market is much more competitive with other operating systems and hardware advancing rapidly. Will the mobile market have an effect on traditional computing, and do you feel as though the Wintel era is coming to an end?
Subject: General Tech, Shows and Expos | June 11, 2011 - 02:28 PM | Scott Michaud
Tagged: john carmack, id, E3
John Carmack was and is one of the biggest faces in videogame engine development since Wolfenstein 3D, Doom, and Quake. He was at E3 to promote his company, iD Software’s, RAGE: their nearest upcoming release. While he was there, PCGamer managed to corner him for a 22 minute interview ranging from RAGE; to the current and future state of PC gaming; to the perceptive effect of input latency and how framerate affects it.
Look at how stable the framerate is!
- Texture resolution and memory limitations on consoles
- Higher end PCs being approximately 10-fold higher performance than the consoles
- Sandy Bridge is finally barely good enough for integrated graphics to be viable GPUs for games
- DirectX and OpenGL APIs hold the PC back, looking forward to new movements to access GPU better
- His interest focuses on the toolset to let the artists do more with less effort
- PC Gaming is still viable but a minority
- Input latency is longer than people expect, sometimes up to 100ms and beyond
- The exciting yet not necessarily crucial nature of newer rendering technologies
John Carmack always has interesting interviews from his very down to Earth and blunt tone. If you have a free half hour and want to hear one of the best game programmers in the world talk about his trade, this is definitely an interview for you.
Subject: General Tech, Displays, Systems | June 11, 2011 - 03:31 AM | Scott Michaud
Tagged: wall tablet, InFocus
InFocus is branding their 55-inch touch-screen TV with Windows 7 embedded as a “Wall Tablet”. The writers down at HotHardware seem to take offense to a 55-inch device being called a tablet and I must agree. My duration working in high schools and acquiring an education degree grew me well acquainted with SMART boards and this product definitely recalls those memories much more vividly than my experience playing around with tablet devices.
The problem with touch screens in schools is that every screen is treated like one thereafter.
(Video from BusinessWire)
It is quite obvious that InFocus spent quite a large amount of time developing their user interface to dress up Windows 7 as a more whiteboard friendly operating system. Their interface has a custom file browser with annotation capabilities, a custom web browser, a digital whiteboard application, and a video conferencing solution that can interface with open protocols such as Google Talk and more proprietary ones such as Cisco. The unit itself has a 720p video camera and a screen resolution of 1920x1080 with multiple touch recognition, something that most (but not all) SMART boards are incapable of.
It is highly unlikely that you will have one of these $6000 devices in your house unless you happen to require it for professional reasons. For those in the education, training, research, or corporate management fields: a device like this could make your life much easier particularly if you were already considering installing a mass of SMART boards for this purpose. They are expected to ship to interested customers in July.