NVIDIA T4 with Ian Buck and Kari Briski

Google Cloud Platform Podcast

Episode | Podcast

Date: Wed, 27 Mar 2019 00:00:00 +0000

<p>Today on the podcast, we speak with <a href="https://twitter.com/ia_buck">Ian Buck</a> and <a href="https://twitter.com/karibriski">Kari Briski</a> of NVIDIA about new updates and achievements in deep learning. Ian begins by telling hosts <a href="https://twitter.com/syntxerror1">Jon</a> and <a href="https://twitter.com/Neurotic">Mark</a> about his first project at NVIDIA, CUDA, and how it has helped expand and pave the way for future projects in super computing, AI, and gaming. CUDA is used extensively in computer vision, speech and audio applications, and machine comprehension, Kari elaborates.</p> <p>NVIDIA recently announced their new Tensor Cores, which maximize their GPUs and make it easier for users to achieve peak performance. Working with the Tensor Cores, TensorFlow AMP is an acceleration into the TensorFlow Framework. It automatically makes the right choices for neural networks and maximizes performance, while still maintaining accuracy, with only a two line change in Tensor Flow script.</p> <p>Just last year, NVIDIA announced their T4 GPU with Google Cloud Platform. This product is designed for inferences, the other side of AI. Because AI is becoming so advanced, complicated, and fast, the GPUs on the inference side have to be able to handle the workload and produce inferences just as quickly. T4 and Google Cloud accomplish this together. Along with T4, NVIDIA has introduced TensorRT, a software framework for AI inference that’s integrated into TensorFlow.</p> <h5 id="ian-buck">Ian Buck</h5> <p><a href="https://twitter.com/ia_buck">Ian Buck</a> is general manager and vice president of Accelerated Computing at NVIDIA. He is responsible for the company’s worldwide datacenter business, including server GPUs and the enabling NVIDIA computing software for AI and HPC used by millions of developers, researchers and scientists. Buck joined NVIDIA in 2004 after completing his PhD in computer science from Stanford University, where he was development lead for Brook, the forerunner to generalized computing on GPUs. He is also the creator of CUDA, which has become the world’s leading platform for accelerated parallel computing. Buck has testified before the U.S. Congress on artificial intelligence and has advised the White House on the topic. Buck also received a BSE degree in computer science from Princeton University.</p> <h5 id="kari-briski">Kari Briski</h5> <p><a href="https://twitter.com/karibriski">Kari Briski</a> is a Senior Director of Accelerated Computing Software Product Management at NVIDIA. Her talents and interests include Deep Learning, Accelerated Computing, Design Thinking, and supporting women in technology. Kari is also a huge Steelers fan.</p> <h5 id="cool-things-of-the-week">Cool things of the week</h5> <ul> <li>Kubernetes 1.14: Production-level support for Windows Nodes, Kubectl Updates, Persistent Local Volumes GA <a href="https://kubernetes.io/blog/2019/03/25/kubernetes-1-14-release-announcement/"> blog</a></li> <li>Stadia <a href="https://www.blog.google/products/stadia/stadia-a-new-way-to-play/"> blog</a></li> <li>How Google Cloud helped Multiplay power a record-breaking Apex Legends launch <a href="https://cloud.google.com/blog/topics/customers/how-google-cloud-helped-multiplay-power-a-record-breaking-apex-legends-launch"> blog</a></li> <li>Massive Entertainment hosts Tom Clancy’s The Division 2 on Google Cloud Platform <a href="https://cloud.google.com/blog/topics/customers/massive-entertainment-hosts-tom-clancys-the-division-2-on-google-cloud-platform"> blog</a></li> </ul> <h5 id="interview">Interview</h5> <ul> <li>NVIDIA <a href="https://www.nvidia.com/en-us/">site</a></li> <li>NVIDIA Catalog <a href="https://ngc.nvidia.com/catalog/landing">site</a></li> <li>CUDA <a href="https://developer.nvidia.com/cuda-toolkit">site</a></li> <li>Tensor Cores <a href="https://developer.nvidia.com/tensor-cores">site</a></li> <li>TensorFlow <a href="https://www.tensorflow.org">sote</a></li> <li>Automatic Mixed Precision for Deep Learning <a href="https://developer.nvidia.com/automatic-mixed-precision">site</a></li> <li>Automatic Mixed Precision for NVIDIA Tensor Core Architecture in TensorFlow <a href="https://devblogs.nvidia.com/nvidia-automatic-mixed-precision-tensorflow/"> blog</a></li> <li>TensorFlow 2.0 on NVIDIA GPU <a href="https://youtu.be/26t8MfP8Fo0">video</a></li> <li>NVIDIA Volta <a href="https://www.nvidia.com/en-us/data-center/volta-gpu-architecture/">site</a></li> <li>NVIDIA T4 <a href="https://www.nvidia.com/en-us/data-center/tesla-t4/">site</a></li> <li>WaveNet <a href="https://deepmind.com/blog/wavenet-generative-model-raw-audio/">blog</a></li> <li>BERT <a href="https://arxiv.org/pdf/1810.04805.pdf">blog</a></li> <li>Compute Engine <a href="https://cloud.google.com/compute/">site</a></li> <li>T4 on GCP <a href="https://cloud.google.com/nvidia/">site</a></li> <li>Webinar On Demand: Accelerate Your AI Models with Automatic Mixed-Precision Training in PyTorch <a href="https://info.nvidia.com/webinar-mixed-precision-with-pytorch-reg-page.html"> site</a></li> <li>PyTorch <a href="https://pytorch.org">site</a></li> <li>NVIDIA TensorRT <a href="https://www.developer.nvidia.com/tensorrt">site</a></li> <li>TensorRT 5.1 <a href="https://news.developer.nvidia.com/speed-up-new-models-with-tensorrt-updates/"> site</a></li> <li>Kubernetes <a href="https://kubernetes.io">site</a></li> <li>Rapids <a href="https://developer.nvidia.com/rapids">site</a></li> <li>NVIDIA GTC <a href="https://www.nvidia.com/en-us/gtc/">site</a></li> <li>Deep Learning Institute <a href="https://www.nvidia.com/en-us/deep-learning-ai/education/">site</a></li> <li>KubeFlow Pipeline Docs <a href="https://www.kubeflow.org/docs/pipelines/pipelines-overview/">site</a></li> <li>KubeFlow Pipelines on GitHub <a href="https://github.com/kubeflow/pipelines">site</a></li> <li>NVIDIA RTX <a href="https://developer.nvidia.com/rtx">site</a></li> </ul> <h5 id="question-of-the-week">Question of the week</h5> <p>Where can we learn more about Stadia?</p> <ul> <li><a href="https://stadia.dev">general info</a></li> <li><a href="https://stadia.dev/apply">developer access</a></li> </ul> <h5 id="where-can-you-find-us-next">Where can you find us next?</h5> <p>Mark will be at <a href="https://cloud.withgoogle.com/next/sf">Cloud NEXT</a>, <a href="http://ecgconf.com">ECGC</a>, and <a href="https://events.google.com/io/">IO</a>.</p> <p>Jon may be going to <a href="https://connect.unity.com/events/UniteShanghai2019">Unite Shanghai</a> and will definitely be at <a href="https://cloud.withgoogle.com/next/sf">Cloud NEXT</a>, <a href="http://ecgconf.com">ECGC</a>, and <a href="https://events.google.com/io/">IO</a>.</p> <p>NVIDIA will be at <a href="https://cloud.withgoogle.com/next/sf">Cloud NEXT</a> and <a href="https://events.linuxfoundation.org/events/kubecon-cloudnativecon-north-america-2019/"> KubeCon</a>, as well as <a href="https://icml.cc">International Conference on Machine Learning</a>, <a href="https://iclr.cc">The International Conference on Learning Representations</a>, and <a href="http://cvpr2019.thecvf.com">CVPR</a></p>