Welcome to DU! The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards. Join the community: Create a free account Support DU (and get rid of ads!): Become a Star Member Latest Breaking News Editorials & Other Articles General Discussion The DU Lounge All Forums Issue Forums Culture Forums Alliance Forums Region Forums Support Forums Help & Search

steve2470

(37,468 posts)
Fri Dec 22, 2017, 02:07 PM Dec 2017

Nvidia's hugely powerful $3,000 Titan V PC GPU is fastest ever

https://www.cnet.com/news/nvidias-3000-titan-v-gpu-offers-massive-power-for-machine-learning/

(a few weeks old but still interesting)

Looking for a way to turn your home computer into a deep-learning AI super-monster? Nvidia has an expensive answer.

The new Titan V GPU promises a crazy amount of processing for deep learning and AI applications. It's nine times more powerful -- at 110 teraflops -- than last year's Titan X, Nvidia's last massive desktop graphics processor aimed at machine learning applications.

The Titan V is based on Nvidia's newer Volta chip architecture, which is also being used in Nvidia's Xavier self-driving car system and for data centers. This is the first desktop-based use of the new chip, and it costs $2,999 or £2,700 (about AU$4,800). Hardly a consumer purchase, but it's meant for people who want a massive dose of machine learning processing. Developers, labs, AI researchers... or anyone who wants an insane supercomputer.

It's available now on Nvidia's store.
4 replies = new reply since forum marked as read
Highlight: NoneDon't highlight anything 5 newestHighlight 5 most recent replies
Nvidia's hugely powerful $3,000 Titan V PC GPU is fastest ever (Original Post) steve2470 Dec 2017 OP
They need to rename 'GPU's htuttle Dec 2017 #1
vector matrix multiplication unit just rolls off the tongue. nt. Voltaire2 Dec 2017 #2
DOe and IBM are installing these new GPUs in the next three lapfog_1 Dec 2017 #3
True, but Graphics Processing Unit doesn't really capture it either, nowadays htuttle Dec 2017 #4

htuttle

(23,738 posts)
1. They need to rename 'GPU's
Fri Dec 22, 2017, 02:12 PM
Dec 2017

So, humans figured out a new use for graphics cards, and now they are used for all sorts of other things.

If you go to Nvidia's homepage for the Titan X, there's not a single mention of 'display monitor' or 'graphics' or anything related to that.

For nostalgia's sake, I'd suggest bringing back the term, "Math Co-Processor".

lapfog_1

(30,225 posts)
3. DOe and IBM are installing these new GPUs in the next three
Fri Dec 22, 2017, 03:50 PM
Dec 2017

fastest Supercomputers in the world... the so-called CORAL project (Collaboration Oak Ridge Argonne Livermore).

https://www.hpcwire.com/2017/12/06/ibm-begins-power9-rollout-backing-doe-google/

Up to 6 of the Nvidia plus 2 PowerPC nextgen CPUs, lots of memory, and NVLink (Nvidia's proprietary replacement for PCIe) AND PCIe Gen4 (next gen PCIe) all in a 2U water chilled (external chilled water required) server.

I think Livermore is installing like 230+ racks of these servers plus next gen Mellanox Infiniband communications between servers. 125+ Pflops at 64 bit floating point. At 16 bit "AI" or "deep learning" should be close to .5 exaflops.

And they are still known as GPUs and not math co-processors because they only do certain floating point math operations at speed...

htuttle

(23,738 posts)
4. True, but Graphics Processing Unit doesn't really capture it either, nowadays
Fri Dec 22, 2017, 11:23 PM
Dec 2017

It's like referring to a rocket as an 'Up Goer'. I'm liking Vector Matrix Multiplication Unit, more and more.

"VMMU" rolls off the tongue just as easily as 'GPU'.

Then we can get back to rating 'video cards' by their RAM, as the Gods intended.

Latest Discussions»Help & Search»Computer Help and Support»Nvidia's hugely powerful ...