Subject: General Tech, Graphics Cards | June 17, 2017 - 09:23 PM | Ken Addison
Tagged: nicehash, mining, cryptocurrency
Over the last several weeks, we have been experimenting with the most recent GPU-shortage-inducing coin mining craze, with Ken's article as a jumping off point. On a recent podcast, I mentioned the idea of running a community coin mining group that would be used as a way for individuals to contribute to PC Perspective. I received several requests for the wallet and setup information to make this happen, so I thought it would be worth while to gather all the necessary links and info in a single location.
We have been running a Patreon campaign for a couple of years now on the site as a way to provide an avenue for those readers and viewers that find PC Perspective a useful resource to the community and directly contribute. It might be because you want to keep the PCPer staff stable, it could be because you use an ad blocker and are looking for a way to even things out, etc. But there are always some that don't have the ability or desire to sign up for a new service so contributing your empty GPU cycles is another option if you want to donate to the PCPer team.
How do you do it? Ken has created a step by step guide below - thanks for your support in this and all of our previous endeavors!
- Bitcoin: 1HHhVWPRpCUst9bDYtLstMdD7o5SzANk1W
- Ethereum: 0xa0294763261aa85eB5f1dA3Ca0f03E1B672EED87
For those of you who may be curious to try out this mining stuff on your personal computer, we would recommend looking into the NiceHash application.
For those of you who haven't read our previous article, NiceHash is a service that connects buyers of GPU mining power to sellers who have spare hardware that they are looking to put to use.
As a warning, if you are planning to mine please be aware of your power consumption. To get a good idea of this, you can look up the TDP of your given graphics card, multiply that wattage by the hours you plan to mine, divide by 1000 to translate from watts to kilowatts, and multiply that by the rate you pay for electricity (this can be found on your power bill in cents per Kilowatt/Hour in the US). (So it's watts*hours*days/1000*kw/hr rate - Thanks CracklingIce)
Given the current rates of value for these cryptocurrencies, power is a small portion of the gross profit made by mining, but it is important to be aware of this before you are presented with a huge power bill that you weren't expecting.
First, download the latest version of the NiceHash miner application from their website.
After your download has finished, extract the ZIP file and load the NiceHashMiner.exe program.
Once the application has been launched and you've accepted the terms of the EULA, the NiceHash Miner will start to download the appropriate mining applications for your given hardware.
Note: during this installation process, your antivirus program might detect malware. These miner executables that are being downloaded are safe, but many antivirius programs flag them as malware because if they are found on your PC without your permission they are a telltale sign of malicious software.
After the installation process is completed, you be brought to the main screen of the application.
From here, choose the server location closest to you, add the Bitcoin address (in this case: 1HHhVWPRpCUst9bDYtLstMdD7o5SzANk1W), and choose a unique worker name (up to 7 characters long).
From here, hit the benchmark button, select the devices you want to mine on (we would recommend GPUs only, CPUs don't earn very much), and hit the Start button.
Once the benchmarking is done, you'll be brought back to the main screen of the application where you can hit the Start button.
Once you hit the start button, a command prompt window will launch where you can see the miner at work (this can be hidden from the NiceHash setting pane), and you can view the stats of your computer in the original NiceHash application window.
And that's it, your computer will now be mining towards the PCPER community pool!
Subject: Graphics Cards | June 14, 2017 - 08:42 PM | Tim Verry
Tagged: zotac, gtx 1080 ti, factory overclocked, gp102, SFF
Zotac recently unveiled a slimmed down GTX 1080 Ti graphics card that uses a dual slot and dual fan cooler with a short PCB. The aptly named Zotac GTX 1080 Ti Mini measures 8.3” (211mm) long and will be the smallest GTX 1080 Ti on the market. Despite the miniaturization, Zotac is still offering a decent factory overclock on the Pascal GPU (but not the memory) with a boost clock of 1620 MHz versus the reference boost clock of 1582.
Zotac uses two 8-pin PCI-E power connectors to drive the card with its GTX 1080 Ti GPU (3584 CUDA cores) and 11GB of GDDR5X memory clocked at 11 GHz. The slimmed down graphics card features a metal backplate, dual shrouded fans, and a heatsink with aluminum fins and five 6mm heat pipes. The card has three DisplayPort 1.4 ports, one HDMI 2.0b port, and one DL-DVI output with the card supporting up to four simultaneous displays.
The Zotac GTX 1080 Ti Mini should enable quite a bit of horsepower in small form factor systems. The graphics card is model number ZT-P10810G-10P and Zotac has it listed on its website. Unfortunately, Zotac is not yet talking pricing or availability for the shortened card.
It appears that overclocking is not out of the question, but I am curious just how far it could be pushed especially in a small case with tight quarters and less airflow.
Subject: General Tech, Graphics Cards | June 13, 2017 - 01:39 PM | Jeremy Hellstrom
Tagged: nvidia, free games, evga, destiny 2
Were you a fan of the original Destiny or simply a fan of free games and happen to be shopping for a new NVIDIA GPU? EVGA have just launched a new giveaway, if you pick up one of their GTX 1080 or 1080 Ti's they will provide you with a code that not only provides you with a free copy of Destiny 2 but also allows you access to the beta.
As usual you need to have an EVGA account so you can register your GPU and so the code can be provided to your account. From there head on over to NVIDIA to redeem the code and patiently await the start of the beta and final release of the game.
June 13th, 2017 - Get Game Ready with EVGA GeForce GTX 10 Series and experience Destiny 2 on PC. For a limited time, buy a select EVGA GeForce GTX 1080 Ti or EVGA GeForce GTX 1080 graphics card and get Destiny 2 at PC Launch and [Early] Access to the PC Beta!
GeForce GTX 10 Series GPUs brings the beautiful world of Destiny 2 to life in stunning 4K. Experience incredibly smooth, tear-free gameplay with NVIDIA G-SYNC™ and share your greatest gameplay moments with NVIDIA ShadowPlay using GeForce Experience.
About Destiny 2:
Humanity's last safe city has fallen to an overwhelming invasion force, led by Ghaul, the imposing commander of the brutal Red Legion. He has stripped the city's Guardians of their power, and forced the survivors to flee. You will venture to mysterious, unexplored worlds of our solar system to discover an arsenal of weapons and devastating new combat abilities. To defeat the Red Legion and confront Ghaul, you must reunite humanity's scattered heroes, stand together, and fight back to reclaim our home.
Learn more and see qualifying EVGA cards at https://www.evga.com/articles/01112/destiny-2-game-ready/
Subject: Graphics Cards | June 13, 2017 - 01:17 PM | Jeremy Hellstrom
Tagged: nvidia, gtx 1080 ti, GTX 1080 Ti GAMING X, msi, Twin Frozr VI, 4k
MSI's latest version of the GeForce GTX 1080 Ti is their GAMING X 4K and has the design features you would expect, Twin Frozr VI, Hi-C CAPs, Super Ferrite Chokes and Japanese Solid Caps. When benchmarking the card [H]ard|OCP saw performance significantly higher than the quoted 1657MHz boost speed, the average was 1935MHz before they overclocked and an impressive 2038MHz for the highest stable in game frequency. They tested both the default and overclocked frequencies against a battery of benchmarks, including the newly released Prey. The card performed admirably at 4k, with many games still performing will with all graphics options at maximum, drop by for a look.
"We review a custom GeForce GTX 1080 Ti based video card with custom cooling and a factory overclock built for overclocking. Can the MSI GeForce GTX 1080 Ti GAMING X truly deliver a consistent enjoyable high-end graphics setting gameplay experience in games at 4K finally? Is a single card viable for current generation gaming at 4K?"
Here are some more Graphics Card articles from around the web:
- Asus ROG Strix GeForce GTX 1080 OC Edition 8GB 11Gbps Video Card Review @ Bjorn3d
- 15-Way NVIDIA/AMD OpenCL GPU Linux Benchmarks Of Ethereum Ethminer @ Phoronix
- XFX RX 460 4GB Heatsink Edition Review @ Bjorn3d
- XFX Rs XXX Edition Rx 570 4GB OC Review @ Bjorn3d
Astute readers of the site might remember the original story we did on Bitcoin mining in 2011, the good ole' days where the concept of the blockchain was new and exciting and mining Bitcoin on a GPU was still plenty viable.
However, that didn't last long, as the race for cash lead people to developing Application Specific Integrated Circuits (ASICs) dedicated solely to Bitcoin mining quickly while sipping power. Use of the expensive ASICs drove the difficulty of mining Bitcoin to the roof and killed any sort of chance of profitability from mere mortals mining cryptocurrency.
Cryptomining saw a resurgence in late 2013 with the popular adoption of alternate cryptocurrencies, specifically Litecoin which was based on the Scrypt algorithm instead of AES-256 like Bitcoin. This meant that the ASIC developed for mining Bitcoin were useless. This is also the period of time that many of you may remember as the "Dogecoin" era, my personal favorite cryptocurrency of all time.
Defenders of these new "altcoins" claimed that Scrypt was different enough that ASICs would never be developed for it, and GPU mining would remain viable for a larger portion of users. As it turns out, the promise of money always wins out, and we soon saw Scrypt ASICs. Once again, the market for GPU mining crashed.
That brings us to today, and what I am calling "Third-wave Cryptomining."
While the mass populous stopped caring about cryptocurrency as a whole, the dedicated group that was left continued to develop altcoins. These different currencies are based on various algorithms and other proofs of works (see technologies like Storj, which use the blockchain for a decentralized Dropbox-like service!).
As you may have predicted, for various reasons that might be difficult to historically quantify, there is another very popular cryptocurrency from this wave of development, Ethereum.
Ethereum is based on the Dagger-Hashimoto algorithm and has a whole host of different quirks that makes it different from other cryptocurrencies. We aren't here to get deep in the woods on the methods behind different blockchain implementations, but if you have some time check out the Ethereum White Paper. It's all very fascinating.
Subject: Graphics Cards | June 8, 2017 - 05:26 PM | Jeremy Hellstrom
Tagged: radeon, Crimson Edition 17.6.1, amd
In the very near future AMD will be releasing an updated driver, focused on improving performance in Prey and DiRT 4.
For DiRT 4 it will enable a Multi GPU profile and up to 30% performance improvement when using 8xMSAA on a Radeon RX 580 8GB compared to the previous release.
Subject: Graphics Cards, Displays | June 6, 2017 - 06:06 PM | Scott Michaud
Tagged: hdr, sdr, nvidia, computex
Dmitry Novoselov of Hardware Canucks saw an NVIDIA SDR vs HDR demo, presumably at Computex based on timing and the intro bumper, and noticed that the SDR monitor looked flat. According to his post in the YouTube comments, he asked NVIDIA to gain access to the monitor settings, and they let him... and he found that the brightness, contrast, and gamma settings were way off. He then performed a factory reset, to test how the manufacturer defaults hold up in the comparison, and did his video based on those results.
I should note that video footage of HDR monitors will not correctly describe what you can see in person. Not only is the camera not HDR, and thus not capable of showing the full range of what the monitor is displaying, but also who knows what the camera’s (and later video processing) exposure and color grading will actually correspond to. That said, he was there and saw it in person, so his eyewitness testimony is definitely valid, but it may or may not focus on qualities that you care about.
Anywho, the test was Mass Effect: Andromeda, which has a native HDR profile. To his taste, he apparently prefers the SDR content in a lot of ways, particularly how the blown out areas behave. He claims that he’s concerned about game-to-game quality, because there will be inconsistency between how one color grading professional chooses to process a scene versus another, but I take issue with that. Even in standard color range, there will always be an art director that decides what looks good and what doesn’t.
They are now given another knob, and it’s an adjustment that the industry is still learning how to deal with, but that’s not a downside to HDR.
Subject: Graphics Cards | June 2, 2017 - 03:02 PM | Jeremy Hellstrom
Tagged: amd, radeon, linux
When Phoronix does a performance round up they do not mess around. Their latest look at the performance of AMD cards on Linux stretches all the way back to the HD 2900XT and encompasses almost every single GPU released between that part and the RX 580, with a pair of Firepro cards and the Fury included as well. For comparative performance numbers you will see 28 NVIDIA cards on these charts, which makes the charts some of the longest you have seen. Drop by to check out the state of AMD performance on Linux in a variety of games as well as synthetic benchmarks.
"It's that time of the year where we see how the open-source AMD Linux graphics driver stack is working on past and present hardware in a large GPU comparison with various OpenGL games and workloads. This year we go from the new Radeon RX 580 all the way back to the Radeon HD 2900XT, looking at how the mature Radeon DRM kernel driver and R600 Gallium3D driver is working for aging ATI/AMD graphics hardware. In total there were 51 graphics cards tested for this comparison of Radeon cards as well as NVIDIA GeForce hardware for reference."
Here are some more Graphics Card articles from around the web:
- PowerColor Red Devl Radeon RX 580 Video Card Review @ Hardware Asylum
- 21-Way NVIDIA Fermi/Kepler/Maxwell/Pascal OpenCL GPU Comparison @ Phoronix
- 28-Way NVIDIA GeForce GPU Comparison On Ubuntu: From GeForce 8 To GeForce 1080 @ Phoronix
- ASUS GTX 1080 ROG Strix OC 11Gbps @ Kitguru
- MSI GTX 1080 Gaming X Plus 8GB @ Kitguru
Subject: Graphics Cards, Mobile | June 2, 2017 - 02:23 AM | Scott Michaud
Tagged: Imagination Technologies, PowerVR, ray tracing, ue4, vulkan
Imagination Technologies has published another video that demonstrates ray tracing with their PowerVR Wizard GPU. The test system, today, is a development card that is running on Ubuntu, and powering Unreal Engine 4. Specifically, it is using UE4’s Vulkan renderer.
The demo highlights two major advantages of ray traced images. The first is that, rather than applying a baked cubemap with screen-space reflections to simulate metallic objects, this demo calculates reflections with secondary rays. From there, it’s just a matter of hooking up the gathered information into the parameters that the shader requires and doing the calculations.
The second advantage is that it can do arbitrary lens effects, like distortion and equirectangular, 360 projections. Rasterization, which projects 3D world coordinates into 2D coordinates on a screen, assumes that edges are still straight, and that causes problems as FoV gets very large, especially full circle. Imagination Technologies acknowledges that workarounds exist, like breaking up the render into six faces of a cube, but the best approximation is casting a ray per pixel and seeing what it hits.
The demo was originally for GDC 2017, back in February, but the videos have just been released.
Subject: Graphics Cards | June 1, 2017 - 02:38 PM | Jeremy Hellstrom
Tagged: plex, live tv
Today Plex announced the addition of a live streaming service to their Plex Pass which will let you stream digital cable channels to any of your devices once you set it up on your account. All you need is a digital tuner such as an NVIDIA SHIELD, a product from Hauppauge, AVerMedia, DVBLogic or even a digital antenna. Hook the device up to the same network your Plex server resides on and start streaming your favourite shows, without paying your cable company rental fees for that set top box. You can check out what channels are available in your area on this Plex page. If you are unfamiliar with Plex or want to read a bit more on the setup you can check out this story at Gizmodo.
"In the latest build of Plex, the server can actually automatically convert the media, resolving this problem. More importantly, you get now watch live TV via a feature called, strangely enough, Live TV."
Here is some more Tech News from around the web:
- Intel Compute Card hands-on @ The Inquirer
- Android execs get technical talking updates, Project Treble, Linux, and more @ Ars Technica
- Enterprise ID management firm OneLogin covfefes to security breach @ The Inquirer
- Nvidia taps Taiwan server industry with HGX reference design for AI systems @ DigiTimes
- Sons of IoT: Bikers hack Jeeps in auto theft spree @ The Register
- Nest leaves competition in the dust with new smart camera @ The Register
- Windows XP Computers Were Mostly Immune To WannaCry @ Slashdot
- Nvidia: Pssst... farmers. Need to get some weeds whacked? @ The Register
- The Dell Computex 2017 Product Launch Coverage @ Tech ARP
- AMD Launches for Threadripper and Radeon RX Vega @ [H]ard|OCP
- The Computex Taipei 2017 Live Coverage (Day 2) @ Tech ARP
- AMD unveils 16-core Ryzen 9 CPUs to take on Intel's Core i9 @ The Inquirer
- The Intel Compute Card, X-Series & Core i9 CPUs & More @ Computex 2017 @ TechARP
- The Computex Taipei 2017 Live Coverage (Day 1) @ TechARP