Log in

goodpods headphones icon

To access all our features

Open the Goodpods app
Close icon
airhacks.fm podcast with adam bien - Accelerating LLMs with TornadoVM: From GPU Kernels to Model Inference

Accelerating LLMs with TornadoVM: From GPU Kernels to Model Inference

05/18/25 • 71 min

airhacks.fm podcast with adam bien
An airhacks.fm conversation with Juan Fumero (@snatverk) about: tornadovm as a Java parallel framework for accelerating data parallelization on GPUs and other hardware, first GPU experiences with ELSA Winner and Voodoo cards, explanation of TornadoVM as a plugin to existing JDKs that uses Graal as a library, TornadoVM's programming model with @parallel and @reduce annotations for parallelizable code, introduction of kernel API for lower-level GPU programming, TornadoVM's ability to dynamically reconfigure and select the best hardware for workloads, implementation of LLM inference acceleration with TornadoVM, challenges in accelerating Llama models on GPUs, introduction of tensor types in TornadoVM to support FP8 and FP16 operations, shared buffer capabilities for GPU memory management, comparison of Java Vector API performance versus GPU acceleration, discussion of model quantization as a potential use case for TornadoVM, exploration of Deep Java Library (DJL) and its ND array implementation, potential standardization of tensor types in Java, integration possibilities with Project Babylon and its Code Reflection capabilities, TornadoVM's execution plans and task graphs for defining accelerated workloads, ability to run on multiple GPUs with different backends simultaneously, potential enterprise applications for LLMs in Java including model distillation for domain-specific models, discussion of Foreign Function & Memory API integration in TornadoVM, performance comparison between different GPU backends like OpenCL and CUDA, collaboration with Intel Level Zero oneAPI and integrated graphics support, future plans for RISC-V support in TornadoVM

Juan Fumero on twitter: @snatverk

plus icon
bookmark
An airhacks.fm conversation with Juan Fumero (@snatverk) about: tornadovm as a Java parallel framework for accelerating data parallelization on GPUs and other hardware, first GPU experiences with ELSA Winner and Voodoo cards, explanation of TornadoVM as a plugin to existing JDKs that uses Graal as a library, TornadoVM's programming model with @parallel and @reduce annotations for parallelizable code, introduction of kernel API for lower-level GPU programming, TornadoVM's ability to dynamically reconfigure and select the best hardware for workloads, implementation of LLM inference acceleration with TornadoVM, challenges in accelerating Llama models on GPUs, introduction of tensor types in TornadoVM to support FP8 and FP16 operations, shared buffer capabilities for GPU memory management, comparison of Java Vector API performance versus GPU acceleration, discussion of model quantization as a potential use case for TornadoVM, exploration of Deep Java Library (DJL) and its ND array implementation, potential standardization of tensor types in Java, integration possibilities with Project Babylon and its Code Reflection capabilities, TornadoVM's execution plans and task graphs for defining accelerated workloads, ability to run on multiple GPUs with different backends simultaneously, potential enterprise applications for LLMs in Java including model distillation for domain-specific models, discussion of Foreign Function & Memory API integration in TornadoVM, performance comparison between different GPU backends like OpenCL and CUDA, collaboration with Intel Level Zero oneAPI and integrated graphics support, future plans for RISC-V support in TornadoVM

Juan Fumero on twitter: @snatverk

Previous Episode

undefined - Run Java with Java

Run Java with Java

An airhacks.fm conversation with Christian Humer (@grashalm_) about: bachelor thesis on a Java bytecode interpreter written in Java, exploration of whether Java could be used as a systems language, benefits of implementing an ecosystem in itself as validation, C1X compiler based on C1 but reimplemented from scratch, concept of sea of nodes for mixing control and data flow, goal to rewrite the entire VM in Java, benefits of using one compiler throughout the stack for compatibility and maintainability, discussion of de-optimization process in JIT compilation, explanation of guards and assumptions in optimized code, three versions of Espresso (Java bytecode interpreter), first version as proof of concept, second version using Truffle with serialized ASTs, third version based on bytecodes with unrolling bytecode loops, explanation of bytecode quickening technique, sandboxing capabilities in GraalVM as replacement for deprecated security manager, isolating untrusted code in separate heaps for security, protection against speculative execution attacks, use case for running AI-generated Java code safely in isolated environments, GraalOS as a minimal operating system for running Java isolates, TRegex as GraalVM's optimized regular expression engine that compiles regex to machine code, bytecode interpreter DSL for generating efficient bytecode interpreters for different languages, memory improvements from using bytecode arrays instead of AST objects, potential future integration of TRegex as a Java API

Christian Humer on twitter: @grashalm_

Episode Comments

Generate a badge

Get a badge for your website that links back to this episode

Select type & size
Open dropdown icon
share badge image

<a href="https://goodpods.com/podcasts/airhacksfm-podcast-with-adam-bien-470050/accelerating-llms-with-tornadovm-from-gpu-kernels-to-model-inference-91414232"> <img src="https://storage.googleapis.com/goodpods-images-bucket/badges/generic-badge-1.svg" alt="listen to accelerating llms with tornadovm: from gpu kernels to model inference on goodpods" style="width: 225px" /> </a>

Copy