|
5 | 5 | <details class=talk><summary><div class="grid"><a href="talks/simd_operations_in_webgpu_for_ml.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/cke8ahza81hlp0731tiz1d295/thumbs/thumb-002.jpeg" alt="Watch SIMD operations in WebGPU for ML" width=200 class="tn"></a><a href="talks/simd_operations_in_webgpu_for_ml.html">SIMD operations in WebGPU for ML</a><span class="summary"> by Mehmet Oguz Derin - 5 min <span></span></span></div><span class=added>Added on 2020-08-26</span></summary><p><a href="talks/simd_operations_in_webgpu_for_ml.html">5 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Mehmet Oguz Derin</dd></dl></details>
|
6 | 6 | <details class=talk><summary><div class="grid"><a href="talks/accelerated_graphics_and_compute_api_for_machine_learning_directml.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckdobv3o075l707723b397arm/thumbs/thumb-004.jpeg" alt="Watch Accelerated graphics and compute API for Machine Learning - DirectML" width=200 class="tn"></a><a href="talks/accelerated_graphics_and_compute_api_for_machine_learning_directml.html">Accelerated graphics and compute API for Machine Learning - DirectML</a><span class="summary"> by Chai Chaoweeraprasit (Microsoft) - 10 min <span></span></span></div></summary><p><a href="talks/accelerated_graphics_and_compute_api_for_machine_learning_directml.html">10 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Chai Chaoweeraprasit (Microsoft)</dd><dd>Chai leads development of machine learning platform at Microsoft</dd><dt>Abstract</dt><dd>DirectML is Microsoft's hardware-accelerated machine learning platform that powers popular frameworks such as TensorFlow and ONNX Runtime. It expands the framework's hardware footprint by enabling high-performance training and inference on any device with DirectX-capable GPU</dd></dl></details>
|
7 | 7 | <details class=talk><summary><div class="grid"><a href="talks/accelerate_ml_inference_on_mobile_devices_with_android_nnapi.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckdo78esr4t0j0772i5r7h1td/thumbs/thumb-001.jpeg" alt="Watch Accelerate ML inference on mobile devices with Android NNAPI" width=200 class="tn"></a><a href="talks/accelerate_ml_inference_on_mobile_devices_with_android_nnapi.html">Accelerate ML inference on mobile devices with Android NNAPI</a><span class="summary"> by Miao Wang (Google) - 7 min <span></span></span></div></summary><p><a href="talks/accelerate_ml_inference_on_mobile_devices_with_android_nnapi.html">7 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Miao Wang (Google)</dd><dd> Software Engineer for Android Neural Networks API</dd><dt>Abstract</dt><dd>The Android Neural Networks API (NNAPI) is an Android C API designed for running computationally intensive operations for machine learning on Android devices. NNAPI is designed to provide a base layer of functionality for higher-level machine learning frameworks, such as TensorFlow Lite and Caffe2, that build and train neural networks. The API is available on all Android devices running Android 8.1 (API level 27) or higher. Based on an app’s requirements and the hardware capabilities on an Android device, NNAPI can efficiently distribute the computation workload across available on-device processors, including dedicated neural network hardware (NPUs and TPUs), graphics processing units (GPUs), and digital signal processors (DSPs).</dd></dl></details>
|
8 |
| -<details><summary><div class="grid"><span>PENDING </span><a>Heterogeneous parallel programming with open standards using oneAPI and Data Parallel C++</a><span class="summary"> by Jeff Hammond (Intel) <span></span></span></div></summary><dl><dt>Speaker</dt><dd>Jeff Hammond (Intel)</dd><dd>Jeff Hammond is a Principal Engineer at Intel where he works on a wide range of high-performance computing topics, including parallel programming models, system architecture and open-source software. He has published more than 60 journal and conference papers on parallel computing, computational chemistry, and linear algebra software. Jeff received his PhD in Physical Chemistry from the University of Chicago.</dd><dt>Abstract</dt><dd>Diversity in computer architecture and the unceasing demand for application performance in data-intensive workloads are never-ending challenges for programmers. This talk will describe Intel’s oneAPI initiative, which is an open ecosystem for heterogeneous computing that supports high-performance data analytics, machine learning and other workloads. A key component of this is Data Parallel C++, which is based on C++17 and Khronos SYCL and supports direct programming of CPU, GPU and FPGA platforms. We will describe how oneAPI and Data Parallel C++ can be used to build high-performance applications for a range of devices.</dd></dl></details> |
| 8 | +<details class=talk><summary><div class="grid"><a href="talks/heterogeneous_parallel_programming_with_open_standards_using_oneapi_and_data_parallel_c_.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckeijprvxa4pq07313u8gw5kd/thumbs/thumb-001.jpeg" alt="Watch Heterogeneous parallel programming with open standards using oneAPI and Data Parallel C++" width=200 class="tn"></a><a href="talks/heterogeneous_parallel_programming_with_open_standards_using_oneapi_and_data_parallel_c_.html">Heterogeneous parallel programming with open standards using oneAPI and Data Parallel C++</a><span class="summary"> by Jeff Hammond (Intel) - 12 min <span></span></span></div><span class=added>Added on 2020-09-01</span></summary><p><a href="talks/heterogeneous_parallel_programming_with_open_standards_using_oneapi_and_data_parallel_c_.html">12 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Jeff Hammond (Intel)</dd><dd>Jeff Hammond is a Principal Engineer at Intel where he works on a wide range of high-performance computing topics, including parallel programming models, system architecture and open-source software. He has published more than 60 journal and conference papers on parallel computing, computational chemistry, and linear algebra software. Jeff received his PhD in Physical Chemistry from the University of Chicago.</dd><dt>Abstract</dt><dd>Diversity in computer architecture and the unceasing demand for application performance in data-intensive workloads are never-ending challenges for programmers. This talk will describe Intel’s oneAPI initiative, which is an open ecosystem for heterogeneous computing that supports high-performance data analytics, machine learning and other workloads. A key component of this is Data Parallel C++, which is based on C++17 and Khronos SYCL and supports direct programming of CPU, GPU and FPGA platforms. We will describe how oneAPI and Data Parallel C++ can be used to build high-performance applications for a range of devices.</dd></dl></details> |
9 | 9 | <details class=talk><summary><div class="grid"><a href="talks/enabling_distributed_dnns_for_the_mobile_web_over_cloud_edge_and_end_devices.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckdobyv0p76q107720jc58j6h/thumbs/thumb-002.jpeg" alt="Watch Enabling Distributed DNNs for the Mobile Web Over Cloud, Edge and End Devices" width=200 class="tn"></a><a href="talks/enabling_distributed_dnns_for_the_mobile_web_over_cloud_edge_and_end_devices.html">Enabling Distributed DNNs for the Mobile Web Over Cloud, Edge and End Devices</a><span class="summary"> by Yakun Huang & Xiuquan Qiao (BPTU) - 9 min <span></span></span></div></summary><p><a href="talks/enabling_distributed_dnns_for_the_mobile_web_over_cloud_edge_and_end_devices.html">9 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Yakun Huang & Xiuquan Qiao (BPTU)</dd><dt>Abstract</dt><dd> This talk introduces two deep learning technologies for the mobile web over cloud, edge and end devices. One is an adaptive DNN execution scheme, which partitions and performs the computation that can be done within the mobile web, reducing the computing pressure of the edge cloud. The other is a lightweight collaborative DNN over cloud, edge and devices, which provides a collaborative mechanism with the edge cloud for accurate compensation.</dd></dl></details>
|
10 | 10 | <details class=talk><summary><div class="grid"><a href="talks/collaborative_learning.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckdobw6x0764l07729krgy57d/thumbs/thumb-001.jpeg" alt="Watch Collaborative Learning" width=200 class="tn"></a><a href="talks/collaborative_learning.html">Collaborative Learning</a><span class="summary"> by Wolfgang Maß (DFKI) - 10 min <span></span></span></div></summary><p><a href="talks/collaborative_learning.html">10 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Wolfgang Maß (DFKI)</dd><dd> Professor at Saarland University and scientific director at DFKI</dd><dt>Abstract</dt><dd> The execution of data analysis services in a browser on devices has recently gained momentum, but the lack of computing resources on devices and data protection regulations are forcing strong constraints. In our talk we will present a browser-based collaborative learning approach for running data analysis services on peer-to-peer networks of devices. Our platform is developed in Javascript, supports modularization of services, model training and usage on devices (tensorflow.js), sensor communication (mqtt), and peer-to-peer communication (WebRTC) with role-based access-control (oauth 2.0).</dd></dl></details>
|
11 | 11 | <details class=talk><summary><div class="grid"><a href="talks/introducing_wasi_nn.html"><img src="https://cjx1uopmt0m4q0667xmnrqpk.blob.core.windows.net/ckdobysfk76ns0772a6u4dbq5/thumbs/thumb-001.jpeg" alt="Watch Introducing WASI-NN" width=200 class="tn"></a><a href="talks/introducing_wasi_nn.html">Introducing WASI-NN</a><span class="summary"> by Mingqiu Sun & Andrew Brown (Intel) - 7 min <span></span></span></div></summary><p><a href="talks/introducing_wasi_nn.html">7 minutes presentation</a></p><dl><dt>Speaker</dt><dd>Mingqiu Sun & Andrew Brown (Intel)</dd><dd> Senior PE at Intel & software engineer at Intel</dd><dt>Abstract</dt><dd> Trained machine learning models are typically deployed on a variety of devices with different architectures and operating systems. WebAssembly provides an ideal portable form of deployment for those models. In this talk, we will introduce the WASI-NN initiative we have started in the WebAssembly System Interface (WASI) community, which would standardize the neural network system interface for WebAssembly programs.</dd></dl></details>
|
|
0 commit comments