hUMA's currently well know trick, a shared memory space which both the CPU and GPU can access without penalty is only the first of its revealed optimizations.  The Register talks today about another way in which this new architecture allows the CPU and GPU equal treatment, standardized task queues and dispatch packets which avoid dealing with a kernel level driver to assign tasks.  With hUMA the GPU is able to shedule tasks for the CPU directly.  That would allow any application that was designed to hUMA standards to have its various tasks assigned to the proper processor without needing extra coding.  This not only makes it cheaper and quicker to design apps but would allow all hUMA apps to take advantage of the specialized abilities of both the CPU and GPU at no cost.

"The upcoming chips will utilise a technique AMD calls Heterogeneous Queuing (hQ). This new approach puts the GPU on an equal footing with the CPU: no longer will the graphics engine have to wait for the central processor to tell it what to do."

Here is some more Tech News from around the web:

Tech Talk