Jake Luciani and Mario Fusco join us to explain the benefits of performing LLM-inference directly in the same JVM as your Java application. They will discuss Jlama's technical aspects, its pros and cons, ongoing work, and future improvements. They will provide a practical example of how Quarkus, LangChain4j, and Jlama simplify creating a pure Java LLM-infused application.