Supercharge your deployments with WASM
Speaker: Rishit Dagli
Machine Learning inference (though what we talk about can be applied to other computationally intensive tasks) is often a computationally intensive task and could greatly benefit from the speed of WebAssembly. What if I told you, that you could improve the performance of your application deployed on Linux containers.
Enter Wasm and Kubernetes.
This talk starts off by introducing the audience to WebAssembly (Wasm) and how they could make use of the speed and security among others of Wasm for their deployments. Another problem we face is that the standard WebAssembly provides very limited access to the native OS and hardware, such as multi-core CPUs, GPU, or TPUS which is not ideal for the kind of systems we target. The talk also shows how one could use the WebAssembly System Interface (WASI) to get security, portability, and native speed for Machine Learning models. The talk then slides into the recent (still preview) support that Azure has launched for WASI node pools to run such a system while also using Krustlet to run Wasm on Kubernetes. To top it off this talk ends with a demo of doing so on Azure.