Nvidia Titan V
Moderator: Site Moderators
Forum rules
Please read the forum rules before posting.
Please read the forum rules before posting.
-
- Site Admin
- Posts: 7926
- Joined: Tue Apr 21, 2009 4:41 pm
- Hardware configuration: Mac Pro 2.8 quad 12 GB smp4
MacBook Pro 2.9 i7 8 GB smp2 - Location: W. MA
Re: Nvidia Titan V
The Nvidia Tesla M2090 is based on a 6 year old Fermi chip, the GF110. Closest comparable consumer level card would be a GTX 580, but the Tesla is running at a slower clock.
iMac 2.8 i7 12 GB smp8, Mac Pro 2.8 quad 12 GB smp6
MacBook Pro 2.9 i7 8 GB smp3
Re: Nvidia Titan V
Update:
Used MSI Afterburner, added 100 Mhz to Core Clock.
Now at 1537 Mhz
Unit 9431 is now at 1165k ppd
Power usage is still at 40%, GPU load 80%,
temp at 60C with 45% fan speed
Used MSI Afterburner, added 100 Mhz to Core Clock.
Now at 1537 Mhz
Unit 9431 is now at 1165k ppd
Power usage is still at 40%, GPU load 80%,
temp at 60C with 45% fan speed
-
- Posts: 2040
- Joined: Sat Dec 01, 2012 3:43 pm
- Hardware configuration: Folding@Home Client 7.6.13 (1 GPU slots)
Windows 7 64bit
Intel Core i5 2500k@4Ghz
Nvidia gtx 1080ti driver 441
Re: Nvidia Titan V
@Aurum: I guess Titans are not worth it because of the high price and similar PPD than gtx 1080ti.
Maybe next summer when gtx 1180/2080 are released these are the cards to go for.
Maybe next summer when gtx 1180/2080 are released these are the cards to go for.
Re: Nvidia Titan V
foldy, I came to the same conclusion. But contrary to economics prices are rising. Must be the crypto miners.
In Science We Trust
-
- Posts: 80
- Joined: Tue Dec 19, 2017 12:19 pm
Re: Nvidia Titan V
Guess I am glad I bought a 7900X so when the Volta GPU's for consumers come out I will not get dinged. Currently running 2 1080TI's windows 10 and getting right at 990000 PPD per card and these stupid little work units. Bigger work units these card rock at 1.3m PPD
-
- Posts: 2040
- Joined: Sat Dec 01, 2012 3:43 pm
- Hardware configuration: Folding@Home Client 7.6.13 (1 GPU slots)
Windows 7 64bit
Intel Core i5 2500k@4Ghz
Nvidia gtx 1080ti driver 441
Re: Nvidia Titan V
And still we have the report of Tesla V100 with 1700k PPD on Linux with 94xx work units. How is this possible compared to Titan V 1200k PPD on Linux?
-
- Site Moderator
- Posts: 6349
- Joined: Sun Dec 02, 2007 10:38 am
- Location: Bordeaux, France
- Contact:
Re: Nvidia Titan V
Different CPUs to feed the GPU ? Different GPU clocks ? More optimized drivers for Tesla (or options not yet activated in the drivers for the newest Titan V) ?
-
- Posts: 2040
- Joined: Sat Dec 01, 2012 3:43 pm
- Hardware configuration: Folding@Home Client 7.6.13 (1 GPU slots)
Windows 7 64bit
Intel Core i5 2500k@4Ghz
Nvidia gtx 1080ti driver 441
Re: Nvidia Titan V
Maybe it would be possible to install a modded nvidia tesla driver on Titan and get better PPD? (This may not be allowed in computer centers)
Re: Nvidia Titan V
I think there's a better chance of getting OpenMM re-optimized to work with the existing Titan driver -- but we're just guessing here. Neither you or I know why performance of the Titan isn't better.
(It also could be that the design of the Titan favours improvements in video performance and power savings over improved in the processing speed/ability of the shaders. Game frame-rate depends on many factors, some of which do not help FAH).
(It also could be that the design of the Titan favours improvements in video performance and power savings over improved in the processing speed/ability of the shaders. Game frame-rate depends on many factors, some of which do not help FAH).
Re: Nvidia Titan V
I don't have a Tesla V100 card to look more deeply into the TitanV lower performance versus the Tesla V100.
Versus Pascal-based cards, the Volta micro architecture is sufficiently different, that some of the performance differences (or lack thereof) of the TitanV vs. 1080Ti could come from a potential need to recompile the binaries (not rewrite Core21, just recompile) - just in case FAH is leveraging CUDA's fatbinaries capability. I can't verify the Tesla V100 numbers side by side with a TitanV in one system, I only know for sure, that with the current settings, PPD's aren't really different vs. a 1080Ti.
I've used the latest CUDA documentation on NVidia's Website to compare the 5 Microarchitectures, or rather the native instruction sets of (Fermi, Kepler, Maxwell, Pascal and Volta) in a single page document (download the PDF). Interestingly, NVidia's documentation indicates that Maxwell and Pascal have identical native instruction sets. Volta has quite a few instructions discontinued from previous generations. They are potentially replaced by other instrutions, potentially more flexible or effective, but compiled code might depend on these differences. For instance, does the Titan driver need to emulate the discontinued instructions in the CPU part of the driver and hence the speed of the CPU is of critical importance for this part of the code. BTW, I used a relatively slow i5-3450 host-CPU. Just an idea, no evidence (yet).
Snippet of the page:
rgds,
Andy
Versus Pascal-based cards, the Volta micro architecture is sufficiently different, that some of the performance differences (or lack thereof) of the TitanV vs. 1080Ti could come from a potential need to recompile the binaries (not rewrite Core21, just recompile) - just in case FAH is leveraging CUDA's fatbinaries capability. I can't verify the Tesla V100 numbers side by side with a TitanV in one system, I only know for sure, that with the current settings, PPD's aren't really different vs. a 1080Ti.
I've used the latest CUDA documentation on NVidia's Website to compare the 5 Microarchitectures, or rather the native instruction sets of (Fermi, Kepler, Maxwell, Pascal and Volta) in a single page document (download the PDF). Interestingly, NVidia's documentation indicates that Maxwell and Pascal have identical native instruction sets. Volta has quite a few instructions discontinued from previous generations. They are potentially replaced by other instrutions, potentially more flexible or effective, but compiled code might depend on these differences. For instance, does the Titan driver need to emulate the discontinued instructions in the CPU part of the driver and hence the speed of the CPU is of critical importance for this part of the code. BTW, I used a relatively slow i5-3450 host-CPU. Just an idea, no evidence (yet).
Snippet of the page:
rgds,
Andy
-
- Posts: 2040
- Joined: Sat Dec 01, 2012 3:43 pm
- Hardware configuration: Folding@Home Client 7.6.13 (1 GPU slots)
Windows 7 64bit
Intel Core i5 2500k@4Ghz
Nvidia gtx 1080ti driver 441
Re: Nvidia Titan V
FP16 is half precision and is not used by FAH but most FP32 single precision and some FP64 double precicion.
There are some 11xxx work units around, maybe there Titan V can hit a new top PPD?
There are some 11xxx work units around, maybe there Titan V can hit a new top PPD?
-
- Site Admin
- Posts: 7926
- Joined: Tue Apr 21, 2009 4:41 pm
- Hardware configuration: Mac Pro 2.8 quad 12 GB smp4
MacBook Pro 2.9 i7 8 GB smp2 - Location: W. MA
Re: Nvidia Titan V
Core_21 does not use CUDA, like the two GPU folding cores before (17 & 18) it uses OpenCL. So recompiling the core is not likely to help.
iMac 2.8 i7 12 GB smp8, Mac Pro 2.8 quad 12 GB smp6
MacBook Pro 2.9 i7 8 GB smp3
Re: Nvidia Titan V
FAHCore_22 has been under development for some time now. (and there's no predicted beta test date or release date). Development plans to release two versions, OpenCL for AMD/ATI and CUDA for NV. I have no doubt that CUDA will improve the performance ofor NV's GPUs but I'm not aware of any testing that has been done to determine whether it will improve Pascal and Titan V equally or unequally. (The same goes for Fermi, Kepler, and Maxwell.) You'll just have to wait and see.
Right now, the objective is to get FAHCore_22 to the point that it can be exposed for beta testing (as well of maximizing overall throughput in any way that is known until then).
Right now, the objective is to get FAHCore_22 to the point that it can be exposed for beta testing (as well of maximizing overall throughput in any way that is known until then).
Posting FAH's log:
How to provide enough info to get helpful support.
How to provide enough info to get helpful support.
Re: Nvidia Titan V
Thanks for the dev update Bruce.
The native instruction set is independent and "beneath" the CUDA or OpenCL layer. The table aggregates native instruction sets of the GPUs based on NVidia's documentation, either framework is compiled to. In the case of CUDA, there is another virtual layer in between (the virtual instruction set PTX). Don't know if OpenCL has a similar architecture.Joe_H wrote:Core_21 does not use CUDA, like the two GPU folding cores before (17 & 18) it uses OpenCL. So recompiling the core is not likely to help.
Re: Nvidia Titan V
True, but from Joe_H's perspective, each recent FAHCore interfaces with either the OpenCL API or the CUDA API.
ATI's OpenCL doesn't pass through a CUDA level. The fact that NVidia's implementation of OpenCL sits on top of CUDA is part of NV's design decision, and rewriting Core_22 to interface directly with CUDA improves efficiency. It's more costly for FAH, though, because they still have to compile a FAHCore that talks to OpenCL so that ATI's GPUs will still be supported -- unless someday ATI decides to pay the license fees for CUDA -- which is never going to happen, IMHO.
ATI's OpenCL doesn't pass through a CUDA level. The fact that NVidia's implementation of OpenCL sits on top of CUDA is part of NV's design decision, and rewriting Core_22 to interface directly with CUDA improves efficiency. It's more costly for FAH, though, because they still have to compile a FAHCore that talks to OpenCL so that ATI's GPUs will still be supported -- unless someday ATI decides to pay the license fees for CUDA -- which is never going to happen, IMHO.
Posting FAH's log:
How to provide enough info to get helpful support.
How to provide enough info to get helpful support.