University of Cambridge > Talks.cam > Technical Talks - Department of Computer Science and Technology  > Mathworks: MATLAB and NVIDIA Docker: A Complete AI Solution, Where You Need It, in an Instant

Mathworks: MATLAB and NVIDIA Docker: A Complete AI Solution, Where You Need It, in an Instant

Add to your list(s) Download to your calendar using vCal

  • UserDr. Jos Martin, Senior Engineering Manager - Parallel Computing
  • ClockMonday 25 November 2019, 13:00-14:30
  • HouseFW26, Computer Laboratory.

If you have a question about this talk, please contact Christine Georgiou.

Food provided

MATLAB ’s deep learning, visualization, and C+/CUDA code generation technology make it a uniquely complete solution for your entire AI workflow. In MATLAB , you can easily manage data, perform complex image and signal processing, prototype and train deep networks, and deploy to your desktop, embedded or cloud environments. Using GPU Coder technology MATLAB generates CUDA kernels that optimize loops and memory access, and C+ that leverages cuDNN and TensorRT, providing the fastest deep network inference of any framework. With MATLAB ’s NVIDIA docker container available through the NVIDIA GPU Cloud, you can now easily access all this AI power, deploy it in your cloud or DGX environment, and get up and running in seconds. In this presentation we will demonstrate a complete end-to-end workflow that starts from ‘docker run’, prototypes and trains a network on a multi-GPU machine in the cloud, and ends with a highly optimized inference engine to deploy to data centers, clouds, and embedded devices.

This talk is part of the Technical Talks - Department of Computer Science and Technology series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity