There are a lot of colloquialisms tossed about such as AI research and machine learning which refer to the work being done designing neural nets by feeding in huge amounts of data to an architecture capable of forming and weighting connections in an attempt to create a system capable of processing that input in a meaningful way. You might be familiar with some of the more famous experiments such as Google's Deep Dream and Wolfram's Language
Image Identification Project. As you might expect this takes a huge amount of computational power and NVIDA has just announced the Tesla M40 accelerator card for training deep neural nets. It is fairly low powered at 50-75W of draw and NVIDIA claims it will be able to deal with five times more simultaneous video streams than previous products. Along with this comes Hyperscale Suite software, specifically designed to work on the new hardware which Jen-Hsun Huang comments on over at The Inquirer.
At the end of the presentation he also mentioned the tiny Jetson TX1 SoC. It has 256-core Maxwell GPU capable of 1TFLOPS, a 64-bit ARM A57 CPU, 4GB of memory and communicates via Ethernet or Wi-Fi all on a card 50x87mm (2×3.4)" in size. It will be available at $300 when released some time early next year.
"Machine learning is the grand computational challenge of our generation. We created the Tesla hyperscale accelerator line to give machine learning a 10X boost. The time and cost savings to data centres will be significant."
Here is some more Tech News from around the web:
- Microsoft Windows Mobile 10: Uphill battle with 'work in progress' @ The Register
- Google gives Chrome for Windows XP a reprieve @ The Inquirer
- ARM's new Cortex-A35: How to fine-tune a CPU for web browsing on bargain smartphones @ The Register
- The Most Powerful DIY Railgun @ Hack a Day
- Apple CEO Tim Cook slags off Microsoft's 'deluded' Surface Book @ The Inquirer
- NVIDIA JTX1: Finally An Exciting 64-bit ARM Board! @ Phoronix
- Cryptowall 4.0: Update makes world's worst ransomware worse still @ The Register
- Sony To End Sales of Betamax Tapes Next Year @ Slashdot
- TSMC 'grand alliance' more powerful than China's 'red supply chain,' says Chang @ DigiTimes
- NitroWare experiences HP's new Australian Customer Experience Centre and Intel 6th Gen ‘Skylake’ PCs
How do they get 1 tflops from
How do they get 1 tflops from 256 Maxwell cores? Gtx 980 has 2048 cores and according to nvidia puts out 4.6 tflops. I doubt this gpu runs at 2x the frequency of gtx 980…
1 TF FP(16 bits) maybe for
1 TF FP(16 bits) maybe for the TX1 kit, but Nvidia is not doing well in the Tablet market and is more car happy at the moment. Denver is dead, and with X1 Nvidia has gone back to using reference ARM cores. Nvidia needs to work towards getting that asynchronous compute fully implemented in GPU hardware. Maybe that’s what Phil Rodgers will be working towards after going to Nvidia from AMD. The VR people are big on asynchronous compute fully implemented in GPU hardware, and for any GPU processing having that true hardware based asynchronous compute will have more advantages and not let any hardware resources remain underutilized for lack of responsive GPU processor thread management.