All About BTC, LTC, ETH mining as well as other alternative crypto currencies
Time for a quick look at the power usage of the GeForce GTX 1070, the reference design from Nvidia that has a TDP of 151W set by the manufacturer, making it pretty energy efficient for the performance it provides for crypto currency mining. We are checking the situation with the different algorithms apart from Ethereum, because we already know that the GTX 1070/1080 is far from great choice for Ethereum mining when using Windows and we are testing the mining for other algorithms with Windows 7.
Looking at the results from the different algorithms the Nvidia GeForce GTX 1070 Founders Edition does seem to be about 30% on average slower (25%-33%) compared to its bigger brother – the Nvidia GeForce GTX 1080 Founders Edition at the stock settings. The situation with the power usage difference between the two cards is also very similar to the performance difference in the tested algorithms. So the next very interesting question is how well the GTX 1070 will overclock and how will the OCed performance compare to a stock GTX 1080 and an overclocked one. We are going to be posting our results from the GTX 1070 overclock soon, so stay tuned for them.
While we were testing and comparing crypto mining performance we have actually noted that the comparison between GTX 1070 and GTX 980 Ti is actually more interesting as you can see from the table with results. As you can clearly see the GTX 1070 is slightly faster than GTX 980 Ti, but power usage wise it is much more efficient. This can give you a better idea on the evolution of performance and power usage between the previous and this generation of Nvidia GPUs. The GTX 1070 is also a more attractive choice for building multi-GPU mining rigs at the moment as compared to the GTX 1080, though you might want to wait a bit more for the non-reference designs to come out that will allow more serious overclocking and thus even better performance.
Time for a quick look at the power usage of the GeForce GTX 1080, the reference design from Nvidia that has a TDP of 180W set by the manufacturer, making it pretty energy efficient for the performance it provides for crypto currency mining. We are checking the situation with the different algorithms apart from Ethereum, because we already know that the GTX 1080 is far from great choice for Ethereum mining.
The GTX 1080 Founders Edition TDP limit of 180 Watt has a power limiter that can give you up to extra 20% increase, moving the slider to increase the maximum available power to be used by the GPU. This can help you get some extra hashrate for the algorithms that max out the 180 Watt TDP of the GPU and as you can see these are actually quite a few. The non reference versions of GTX 1080 should come with higher TDP limits and even higher factory clocks, so they should be able to achieve even more performance as they will be able to keep higher boost frequency of the GPU while mining. Then again you might also want to lower the power limit in order to reduce the power usage and get an even better hashrate per Watt of power used, but for that you might have to also play with the voltage of the GPU for best results and also the clocks.
In the table above you can see that Neoscrypt and Lyra2Re are far from maxing out the TDP of the GTX 1080, the reason for that is that they do not perform that well in terms of hashrate, so some optimizations for them might help give you better performance. Other than that the power usage is high and the hashrate you get is seriously increased as compared to the previous generation of Nvidia GPUs, so performance per Watt for the GTX 1080 is actually great in most algorithms for crypto mining.
Lately there is much talk about power efficiency of various mining algorithms and with the summer here people with GPU mining rigs are looking for algorithms that use less power and thus the video cards run cooler and quieter. We are starting a series of tests with GeForce GTX 750 Ti GPU first and then we are also going to move to other popular video cards for mining crypto currencies such as the Radeon R9 280X for example.
On the photo above you can see the power usage of the GTX 750 Ti video card in idle as well as the idle power usage of the whole system we are using for testing; below you can find the specifications of the hardware. Note that one of the power meters measures only the power usage of the video card (the power meter is attached to the power line going to the card directly and all power going to it passes through the meter, so it does not take into account the PSU power efficiency) and the other one is for the whole system measured at the wall (the actual full power consumption) taking into account the efficiency of the power supply (extra power wasted as heat during the conversion).
The systems we are using for the tests include:
– Palit GeForce GTX 750 Ti StormX OC 2GB video card
– Intel i3-4130 dual-core CPU at 3.4 GHz
– Asus H81M-A Motherboard
– 2x 4GB A-DATA DDR3 1600 MHz Memory
– 1TB Seagate 7200 RPM Hard drive
– 500W Cooler Master Power Supply
We have used ccMiner for our tests, the latest fork with Fresh algorithm support and we have measured the power usage of the GPU only as well as of the whole system with all of the supported algorithms by that particular version of ccMiner. Do note that if mining for Scrypt for example you will be getting higher power usage, but this is already pretty pointless to be done with GPUs with so many Scrypt ASIC miners already deployed. The results we’ve seen on the GTX 750 Ti are pretty interesting; it seems that the most power efficient algorithms are Fugue256 and HEFTY1 with the new Fresh algorithm following close by with the same power usage as Qubit. The worst performing crypto algorithms on GTX 750 Ti are the Groestl-based ones and the X-ones are pretty much in the middle. Do note however that these are the results measured on GTX 750 Ti, the situation with AMD with the same algorithms may differ significantly and we do plan to run some tests to check the situation there as well, so stay tuned for more very soon, probably tomorrow.
Since there were some questions and people doubting our measurements, we have repeated the tests with another power meter connected to measure the power going only to the video card and the results are pretty much the same as with the previous meter in terms of power usage as you can see on the photo above. Do note that the Palit GeForce GTX 750 Ti video card that we have used for testing does not have an external PCI-E power connector available, so all of the power going to the video card is from the PCI-E slot. So in order to measure the exact power used by the video card we have used a powered PCI Express x1-x16 USB 3.0 Extender. This extender does not use USB 3.0 interface, just a USB 3.0 data cable for the transmission of data between the PCI-E slot on the motherboard and the video card (no power is transmitted over that cable). Instead the power provided to the video card all goes through the 4-pin Molex power connector on the extender’s board. Also do note that the power measured is coming directly from the power supply, so this measurement for the power usage of the GPU does not take into account the power efficiency of the power supply (loses of power during the conversion from 110V/220V to 12V) and depending on the power supply there will be about 10-20% of extra power lost as heat during the conversion. This power is measured by the second power meter that does measure the full system’s power consumption at the power socket however.