Weekly talks and fireside chats about everything that has to do with the new space emerging around DevOps for Machine Learning aka MLOps aka Machine Learning Operations.
The podcast MLOps.community is created by Demetrios Brinkmann. The podcast and the artwork on this page are embedded on this page using the public podcast feed (RSS).
Jineet Doshi is a Staff Data Scientist at Intuit and an AI Lead with a strong background in Computer Science. With over 6 years of relevant experience, he has a proven track record of building end-to-end machine learning models that significantly improve business metrics, from reducing fraud to saving millions of dollars. Holistic Evaluation of Generative AI Systems // MLOps Podcast #280 with Jineet Doshi, Staff AI Scientist or AI Lead at Intuit. // Abstract Evaluating LLMs is essential in establishing trust before deploying them to production. Even post deployment, evaluation is essential to ensure LLM outputs meet expectations, making it a foundational part of LLMOps. However, evaluating LLMs remains an open problem. Unlike traditional machine learning models, LLMs can perform a wide variety of tasks such as writing poems, Q&A, summarization etc. This leads to the question how do you evaluate a system with such broad intelligence capabilities? This talk covers the various approaches for evaluating LLMs such as classic NLP techniques, red teaming and newer ones like using LLMs as a judge, along with the pros and cons of each. The talk includes evaluation of complex GenAI systems like RAG and Agents. It also covers evaluating LLMs for safety and security and the need to have a holistic approach for evaluating these very capable models. // Bio Jineet Doshi is an award winning AI Lead and Engineer with over 7 years of experience. He has a proven track record of leading successful AI projects and building machine learning models from design to production across various domains, which have impacted millions of customers and have significantly improved business metrics, leading to millions of dollars of impact. He is currently an AI Lead at Intuit where he is one of the architects and developers of their Generative AI platform, which is serving Generative AI experiences for more than 100 million customers around the world. Jineet is also a guest lecturer at Stanford University as part of their building LLM Applications class. He is on the Advisory Board of University of San Francisco’s AI Program. He holds multiple patents in the field, is on the steering committee of MLOps World Conference and has also co chaired workshops at top AI conferences like KDD. He holds a Masters degree from Carnegie Mellon university. // MLOps Swag/Merch https://shop.mlops.community/ // Related Links Website: https://www.intuit.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jineet on LinkedIn: https://www.linkedin.com/in/jineetdoshi/
Robert Caulk is responsible for directing software development, enabling research, coordinating company projects, quality control, proposing external collaborations, and securing funding. He believes firmly in open-source, having spent 12 years accruing over 1000 academic citations building open-source software in domains such as machine learning, image analysis, and coupled physical processes. He received his Ph.D. from Université Grenoble Alpes, France, in computational mechanics. Unleashing Unconstrained News Knowledge Graphs to Combat Misinformation // MLOps Podcast #279 with Robert Caulk, Founder of Emergent Methods. // Abstract Indexing hundreds of thousands of news articles per day into a knowledge graph (KG) was previously impossible due to the strict requirement that high-level reasoning, general world knowledge, and full-text context *must* be present for proper KG construction. The latest tools now enable such general world knowledge and reasoning to be applied cost effectively to high-volumes of news articles. Beyond the low cost of processing these news articles, these tools are also opening up a new, controversial, approach to KG building - unconstrained KGs. We discuss the construction and exploration of the largest news-knowledge-graph on the planet - hosted on an endpoint at AskNews.app. During talk we aim to highlight some of the sacrifices and benefits that go hand-in-hand with using the infamous unconstrained KG approach. We conclude the talk by explaining how knowledge graphs like these help to mitigate misinformation. We provide some examples of how our clients are using this graph, such as generating sports forecasts, generating better social media posts, generating regional security alerts, and combating human trafficking. // Bio Robert is the founder of Emergent Methods, where he directs research and software development for large-scale applications. He is currently overseeing the structuring of hundreds of thousands of news articles per day in order to build the best news retrieval API in the world: https://asknews.app. // MLOps Swag/Merch
https://shop.mlops.community/ // Related Links Website: https://emergentmethods.ai
News Retrieval API: https://asknews.app --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Rob on LinkedIn: https://www.linkedin.com/in/rcaulk/
Timestamps: [00:00] Rob's preferred coffee [00:05] Takeaways [00:55] Please like, share, leave a review, and subscribe to our MLOps channels! [01:00] Join our Local Organizer Carousel! [02:15] Knowledge Graphs and ontology [07:43] Ontology vs Noun Approach [12:46] Ephemeral tools for efficiency [17:26] Oracle to PostgreSQL migration [22:20] MEM Graph life cycle [29:14] Knowledge Graph Investigation Insights [33:37] Fine-tuning and distillation of LLMs [39:28] DAG workflow and quality control [46:23] Crawling nodes with Phi 3 Llama [50:05] AI pricing risks and strategies [56:14] Data labeling and poisoning [58:34] API costs vs News latency [1:02:10] Product focus and value [1:04:52] Ensuring reliable information [1:11:01] Podcast transcripts as News [1:13:08] Ontology trade-offs explained [1:15:00] Wrap up
Guanhua Wang is a Senior Researcher in DeepSpeed Team at Microsoft. Before Microsoft, Guanhua earned his Computer Science PhD from UC Berkeley. Domino: Communication-Free LLM Training Engine // MLOps Podcast #278 with Guanhua "Alex" Wang, Senior Researcher at Microsoft. // Abstract Given the popularity of generative AI, Large Language Models (LLMs) often consume hundreds or thousands of GPUs to parallelize and accelerate the training process. Communication overhead becomes more pronounced when training LLMs at scale. To eliminate communication overhead in distributed LLM training, we propose Domino, which provides a generic scheme to hide communication behind computation. By breaking the data dependency of a single batch training into smaller independent pieces, Domino pipelines these independent pieces of training and provides a generic strategy of fine-grained communication and computation overlapping. Extensive results show that compared with Megatron-LM, Domino achieves up to 1.3x speedup for LLM training on Nvidia DGX-H100 GPUs. // Bio Guanhua Wang is a Senior Researcher in the DeepSpeed team at Microsoft. His research focuses on large-scale LLM training and serving. Previously, he led the ZeRO++ project at Microsoft which helped reduce over half of model training time inside Microsoft and Linkedin. He also led and was a major contributor to Microsoft Phi-3 model training. He holds a CS PhD from UC Berkeley advised by Prof Ion Stoica. // MLOps Swag/Merch https://shop.mlops.community/ // Related Links Website: https://guanhuawang.github.io/ DeepSpeed hiring: https://www.microsoft.com/en-us/research/project/deepspeed/opportunities/
Large Model Training and Inference with DeepSpeed // Samyam Rajbhandari // LLMs in Prod Conference: https://youtu.be/cntxC3g22oU --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Guanhua on LinkedIn: https://www.linkedin.com/in/guanhua-wang/ Timestamps: [00:00] Guanhua's preferred coffee [00:17] Takeaways [01:36] Please like, share, leave a review, and subscribe to our MLOps channels! [01:47] Phi model explanation [06:29] Small Language Models optimization challenges [07:29] DeepSpeed overview and benefits [10:58] Crazy unimplemented crazy AI ideas [17:15] Post training vs QAT [19:44] Quantization over distillation [24:15] Using Lauras [27:04] LLM scaling sweet spot [28:28] Quantization techniques [32:38] Domino overview [38:02] Training performance benchmark [42:44] Data dependency-breaking strategies [49:14] Wrap up
Thanks to the High Signal Podcast by Delphina: https://go.mlops.community/HighSignalPodcast
Aditya Naganath is an experienced investor currently working with Kleiner Perkins. He has a passion for connecting with people over coffee and discussing various topics related to tech, products, ideas, and markets.
AI's Next Frontier // MLOps Podcast #277 with Aditya Naganath, Principal at Kleiner Perkins.
// Abstract
LLMs have ushered in an unmistakable supercycle in the world of technology. The low-hanging use cases have largely been picked off. The next frontier will be AI coworkers who sit alongside knowledge workers, doing work side by side. At the infrastructure level, one of the most important primitives invented by man - the data center, is being fundamentally rethought in this new wave.
// Bio
Aditya Naganath joined Kleiner Perkins’ investment team in 2022 with a focus on artificial intelligence, enterprise software applications, infrastructure and security. Prior to joining Kleiner Perkins, Aditya was a product manager at Google focusing on growth initiatives for the next billion users team. He previously was a technical lead at Palantir Technologies and formerly held software engineering roles at Twitter and Nextdoor, where he was a Kleiner Perkins fellow. Aditya earned a patent during his time at Twitter for a technical analytics product he co-created.
Originally from Mumbai India, Aditya graduated magna cum laude from Columbia University with a bachelor’s degree in Computer Science, and an MBA from Stanford University. Outside of work, you can find him playing guitar with a hard rock band, competing in chess or on the squash courts, and fostering puppies. He is also an avid poker player.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Faith's Hymn by Beautiful Chorus: https://open.spotify.com/track/1bDv6grQB5ohVFI8UDGvKK?si=4b00752eaa96413b Substack: https://adityanaganath.substack.com/?utm_source=substack&utm_medium=web&utm_campaign=substack_profile
With thanks to the High Signal Podcast by Delphina: https://go.mlops.community/HighSignalPodcast
Building the Future of AI in Software Development // Varun Mohan // MLOps Podcast #195 - https://youtu.be/1DJKq8StuTo
Do Re MI for Training Metrics: Start at the Beginning // Todd Underwood // AIQCON - https://youtu.be/DxyOlRdCofo
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Aditya on LinkedIn: https://www.linkedin.com/in/aditya-naganath/
Dr Vincent Moens is an Applied Machine Learning Research Scientist at Meta and an author of TorchRL and TensorDict in Pytorch. PyTorch for Control Systems and Decision Making // MLOps Podcast #276 with Vincent Moens, Research Engineer at Meta. // Abstract PyTorch is widely adopted across the machine learning community for its flexibility and ease of use in applications such as computer vision and natural language processing. However, supporting reinforcement learning, decision-making, and control communities is equally crucial, as these fields drive innovation in areas like robotics, autonomous systems, and game-playing. This podcast explores the intersection of PyTorch and these fields, covering practical tips and tricks for working with PyTorch, an in-depth look at TorchRL, and discussions on debugging techniques, optimization strategies, and testing frameworks. By examining these topics, listeners will understand how to effectively use PyTorch for control systems and decision-making applications. // Bio Vincent Moens is a research engineer on the PyTorch core team at Meta, based in London. As the maintainer of TorchRL (https://github.com/pytorch/rl) and TensorDict (https://github.com/pytorch/tensordict), Vincent plays a key role in supporting the decision-making community within the PyTorch ecosystem. Alongside his technical role in the PyTorch community, Vincent also actively contributes to AI-related research projects. Before joining Meta, Vincent worked as an ML researcher at Huawei and AIG. Vincent holds a Medical Degree and a PhD in Computational Neuroscience. // MLOps Swag/Merch https://shop.mlops.community/ // Related Links
Musical recommendation: https://open.spotify.com/artist/1Uff91EOsvd99rtAupatMP?si=jVkoFiq8Tmq0fqK_OIEglg Website: github.com/vmoens TorchRL: https://github.com/pytorch/rl TensorDict: https://github.com/pytorch/tensordict LinkedIn post: https://www.linkedin.com/posts/vincent-moens-9bb91972_join-the-tensordict-discord-server-activity-7189297643322253312-Wo9J?utm_source=share&utm_medium=member_desktop --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vincent on LinkedIn: https://www.linkedin.com/in/mvi/
Matt Van Itallie is the founder and CEO of Sema. Prior to this, they were the Vice President of Customer Support and Customer Operations at Social Solutions. AI-Driven Code: Navigating Due Diligence & Transparency in MLOps // MLOps Podcast #275 with Matt van Itallie, Founder and CEO of Sema. // Abstract Matt Van Itallie, founder and CEO of Sema, discusses how comprehensive codebase evaluations play a crucial role in MLOps and technical due diligence. He highlights the impact of Generative AI on code transparency and explains the Generative AI Bill of Materials (GBOM), which helps identify and manage risks in AI-generated code. This talk offers practical insights for technical and non-technical audiences, showing how proper diligence can enhance value and mitigate risks in machine learning operations. // Bio Matt Van Itallie is the Founder and CEO of Sema. He and his team have developed Comprehensive Codebase Scans, the most thorough and easily understandable assessment of a codebase and engineering organization. These scans are crucial for private equity and venture capital firms looking to make informed investment decisions. Sema has evaluated code within organizations that have a collective value of over $1 trillion. In 2023, Sema served 7 of the 9 largest global investors, along with market-leading strategic investors, private equity, and venture capital firms, providing them with critical insights. In addition, Sema is at the forefront of Generative AI Code Transparency, which measures how much code created by GenAI is in a codebase. They are the inventors behind the Generative AI Bill of Materials (GBOM), an essential resource for investors to understand and mitigate risks associated with AI-generated code. Before founding Sema, Matt was a Private Equity operating executive and a management consultant at McKinsey. He graduated from Harvard Law School and has had some interesting adventures, like hiking a third of the Appalachian Trail and biking from Boston to Seattle. Full bio: https://alistar.fm/bio/matt-van-itallie // MLOps Swag/Merch https://shop.mlops.community/ // Related Links Website: https://en.m.wikipedia.org/wiki/Michael_Gschwind --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Matt on LinkedIn: https://www.linkedin.com/in/mvi/
Dr. Michael Gschwind is a Director / Principal Engineer for PyTorch at Meta Platforms. At Meta, he led the rollout of GPU Inference for production services. // MLOps Podcast #274 with Michael Gschwind, Software Engineer, Software Executive at Meta Platforms. // Abstract Explore the role in boosting model performance, on-device AI processing, and collaborations with tech giants like ARM and Apple. Michael shares his journey from gaming console accelerators to AI, emphasizing the power of community and innovation in driving advancements. // Bio Dr. Michael Gschwind is a Director / Principal Engineer for PyTorch at Meta Platforms. At Meta, he led the rollout of GPU Inference for production services. He led the development of MultiRay and Textray, the first deployment of LLMs at a scale exceeding a trillion queries per day shortly after its rollout. He created the strategy and led the implementation of PyTorch donation optimization with Better Transformers and Accelerated Transformers, bringing Flash Attention, PT2 compilation, and ExecuTorch into the mainstream for LLMs and GenAI models. Most recently, he led the enablement of large language models on-device AI with mobile and edge devices. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://en.m.wikipedia.org/wiki/Michael_Gschwind --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Michael on LinkedIn: https://www.linkedin.com/in/michael-gschwind-3704222/?utm_source=share&utm_campaign=share_via&utm_content=profile&utm_medium=ios_app
Timestamps: [00:00] Michael's preferred coffee [00:21] Takeaways [01:59] Please like, share, leave a review, and subscribe to our MLOps channels! [02:10] Gaming to AI Accelerators [11:34] Torch Chat goals [18:53] Pytorch benchmarking and competitiveness [21:28] Optimizing MLOps models [24:52] GPU optimization tips [29:36] Cloud vs On-device AI [38:22] Abstraction across devices [42:29] PyTorch developer experience [45:33] AI and MLOps-related antipatterns [48:33] When to optimize [53:26] Efficient edge AI models [56:57] Wrap up
//Abstract In this segment, the Panel will dive into the evolving landscape of AI, where large language models (LLMs) power the next wave of intelligent agents. In this engaging panel, leading investors Meera (Redpoint), George (Sequoia), and Sandeep (Prosus Ventures) discuss the promise and pitfalls of AI in production. From transformative industry applications to the challenges of scalability, costs, and shifting business models, this session unpacks the metrics and insights shaping GenAI's future. Whether you're excited about AI's potential or wary of its complexities, this is a must-watch for anyone exploring the cutting edge of tech investment. //Bio Host: Paul van der Boor Senior Director Data Science @ Prosus Group Sandeep Bakshi Head of Investments, Europe @ Prosus Meera Clark Principal @ Redpoint Ventures George Robson Partner @ Sequoia Capital A Prosus | MLOps Community Production
Luke Marsden, is a passionate technology leader. Experienced in consultant, CEO, CTO, tech lead, product, sales, and engineering roles. Proven ability to conceive and execute a product vision from strategy to implementation, while iterating on product-market fit. We Can All Be AI Engineers and We Can Do It with Open Source Models // MLOps Podcast #273 with Luke Marsden, CEO of HelixML. // Abstract In this podcast episode, Luke Marsden explores practical approaches to building Generative AI applications using open-source models and modern tools. Through real-world examples, Luke breaks down the key components of GenAI development, from model selection to knowledge and API integrations, while highlighting the data privacy advantages of open-source solutions. // Bio Hacker & entrepreneur. Founder at helix.ml. Career spanning DevOps, MLOps, and now LLMOps. Working on bringing business value to local, open-source LLMs. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://helix.ml About open source AI: https://blog.helix.ml/p/the-open-source-ai-revolution
Ratatat Cream on Chrome: https://open.spotify.com/track/3s25iX3minD5jORW4KpANZ?si=719b715154f64a5f --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Luke on LinkedIn: https://www.linkedin.com/in/luke-marsden-71b3789/
//Abstract This panel speaks about the diverse landscape of AI agents, focusing on how they integrate voice interfaces, GUIs, and small language models to enhance user experiences. They'll also examine the roles of these agents in various industries, highlighting their impact on productivity, creativity, and user experience and how these empower developers to build better solutions while addressing challenges like ensuring consistent performance and reliability across different modalities when deploying AI agents in production. //Bio Host: Diego Oppenheimer Co-founder @ Guardrails AI Jazmia Henry Founder and CEO @ Iso AI Rogerio Bonatti Researcher @ Microsoft Julia Kroll Applied Engineer @ Deepgram Joshua Alphonse Director of Developer Relations @ PremAI A Prosus | MLOps Community Production
Lauren Kaplan is a sociologist and writer. She earned her PhD in Sociology at Goethe University Frankfurt and worked as a researcher at the University of Oxford and UC Berkeley. The Impact of UX Research in the AI Space // MLOps Podcast #272 with Lauren Kaplan, Sr UX Researcher. // Abstract In this MLOps Community podcast episode, Demetrios and UX researcher Lauren Kaplan explore how UX research can transform AI and ML projects by aligning insights with business goals and enhancing user and developer experiences. Kaplan emphasizes the importance of stakeholder alignment, proactive communication, and interdisciplinary collaboration, especially in adapting company culture post-pandemic. They discuss UX’s growing relevance in AI, challenges like bias, and the use of AI in research, underscoring the strategic value of UX in driving innovation and user satisfaction in tech. // Bio Lauren is a sociologist and writer. She earned her PhD in Sociology at Goethe University Frankfurt and worked as a researcher at the University of Oxford and UC Berkeley. Passionate about homelessness and Al, Lauren joined UCSF and later Meta. Lauren recently led UX research at a global Al chip startup and is currently seeking new opportunities to further her work in UX research and AI. At Meta, Lauren led UX research for 1) Privacy-Preserving ML and 2) PyTorch. Lauren has worked on NLP projects such as Word2Vec analysis of historical HIV/AIDS documents presented at TextXD, UC Berkeley 2019. Lauren is passionate about understanding technology and advocating for the people who create and consume Al. Lauren has published over 30 peer-reviewed research articles in domains including psychology, medicine, sociology, and more.” // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Podcast on AI UX https://open.substack.com/pub/aistudios/p/how-to-do-user-research-for-ai-products?r=7hrv8&utm_medium=ios 2024 State of AI Infra at Scale Research Report https://ai-infrastructure.org/wp-content/uploads/2024/03/The-State-of-AI-Infrastructure-at-Scale-2024.pdf Privacy-Preserving ML UX Public Article https://www.ttclabs.net/research/how-to-help-people-understand-privacy-enhancing-technologies Homelessness research and more: https://scholar.google.com/citations?user=24zqlwkAAAAJ&hl=en Agents in Production: https://home.mlops.community/public/events/aiagentsinprod
Mk.gee Si (Bonus Track): https://open.spotify.com/track/1rukW2Wxnb3GGlY0uDWIWB?si=4d5b0987ad55444a --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Lauren on LinkedIn: https://www.linkedin.com/in/laurenmichellekaplan?utm_source=share&utm_campaign=share_via&utm_content=profile&utm_medium=ios_app
Dr. Petar Tsankov is a researcher and entrepreneur in the field of Computer Science and Artificial Intelligence (AI). EU AI Act - Navigating New Legislation // MLOps Podcast #271 with Petar Tsankov, Co-Founder and CEO of LatticeFlow AI. Big thanks to LatticeFlow for sponsoring this episode! // Abstract Dive into AI risk and compliance. Petar Tsankov, a leader in AI safety, talks about turning complex regulations into clear technical requirements and the importance of benchmarks in AI compliance, especially with the EU AI Act. We explore his work with big AI players and the EU on safer, compliant models, covering topics from multimodal AI to managing AI risks. He also shares insights on "Comply," an open-source tool for checking AI models against EU standards, making compliance simpler for AI developers. A must-listen for those tackling AI regulation and safety. // Bio Co-founder & CEO at LatticeFlow AI, building the world's first product enabling organizations to build performant, safe, and trustworthy AI systems. Before starting LatticeFlow AI, Petar was a senior researcher at ETH Zurich working on the security and reliability of modern systems, including deep learning models, smart contracts, and programmable networks. Petar have co-created multiple publicly available security and reliability systems that are regularly used: = ERAN, the world's first scalable verifier for deep neural networks: https://github.com/eth-sri/eran = VerX, the world's first fully automated verifier for smart contracts: https://verx.ch = Securify, the first scalable security scanner for Ethereum smart contracts: https://securify.ch = DeGuard, de-obfuscates Android binaries: http://apk-deguard.com = SyNET, the first scalable network-wide configuration synthesis tool: https://synet.ethz.ch Petar also co-founded ChainSecurity, an ETH spin-off that within 2 years became a leader in formal smart contract audits and was acquired by PwC Switzerland in 2020. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://latticeflow.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Petar on LinkedIn: https://www.linkedin.com/in/petartsankov/
Bernie Wu is VP of Business Development for MemVerge. He has 25+ years of experience as a senior executive for data center hardware and software infrastructure companies including companies such as Conner/Seagate, Cheyenne Software, Trend Micro, FalconStor, Levyx, and MetalSoft. Boosting LLM/RAG Workflows & Scheduling w/ Composable Memory and Checkpointing // MLOps Podcast #270 with Bernie Wu, VP Strategic Partnerships/Business Development of MemVerge. // Abstract Limited memory capacity hinders the performance and potential of research and production environments utilizing Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) techniques. This discussion explores how leveraging industry-standard CXL memory can be configured as a secondary, composable memory tier to alleviate this constraint. We will highlight some recent work we’ve done in integrating of this novel class of memory into LLM/RAG/vector database frameworks and workflows. Disaggregated shared memory is envisioned to offer high performance, low latency caches for model/pipeline checkpoints of LLM models, KV caches during distributed inferencing, LORA adaptors, and in-process data for heterogeneous CPU/GPU workflows. We expect to showcase these types of use cases in the coming months. // Bio Bernie is VP of Strategic Partnerships/Business Development for MemVerge. His focus has been building partnerships in the AI/ML, Kubernetes, and CXL memory ecosystems. He has 25+ years of experience as a senior executive for data center hardware and software infrastructure companies including companies such as Conner/Seagate, Cheyenne Software, Trend Micro, FalconStor, Levyx, and MetalSoft. He is also on the Board of Directors for Cirrus Data Solutions. Bernie has a BS/MS in Engineering from UC Berkeley and an MBA from UCLA. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.memverge.com Accelerating Data Retrieval in Retrieval Augmentation Generation (RAG) Pipelines using CXL: https://memverge.com/accelerating-data-retrieval-in-rag-pipelines-using-cxl/ Do Re MI for Training Metrics: Start at the Beginning // Todd Underwood // AIQCON: https://youtu.be/DxyOlRdCofo Handling Multi-Terabyte LLM Checkpoints // Simon Karasik // MLOps Podcast #228: https://youtu.be/6MY-IgqiTpg
Compute Express Link (CXL) FPGA IP: https://www.intel.com/content/www/us/en/products/details/fpga/intellectual-property/interface-protocols/cxl-ip.htmlUltra Ethernet Consortium: https://ultraethernet.org/
Unified Acceleration (UXL) Foundation: https://www.intel.com/content/www/us/en/developer/articles/news/unified-acceleration-uxl-foundation.html
RoCE networks for distributed AI training at scale: https://engineering.fb.com/2024/08/05/data-center-engineering/roce-network-distributed-ai-training-at-scale/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Bernie on LinkedIn: https://www.linkedin.com/in/berniewu/
Timestamps: [00:00] Bernie's preferred coffee [00:11] Takeaways [01:37] First principles thinking focus [05:02] Memory Abundance Concept Discussion [06:45] Managing load spikes [09:38] GPU checkpointing challenges [16:29] Distributed memory problem solving [18:27] Composable and Virtual Memory [21:49] Interactive chat annotation [23:46] Memory elasticity in AI [27:33] GPU networking tests [29:12] GPU Scheduling workflow optimization [32:18] Kubernetes Extensions and Tools [37:14] GPU bottleneck analysis [42:04] Economical memory strategies [45:14] Elastic memory management strategies [47:57] Problem solving approach [50:15] AI infrastructure elasticity evolution [52:33] RDMA and RoCE explained [54:14] Wrap up
Gideon Mendels is the Chief Executive Officer at Comet, the leading solution for managing machine learning workflows. How to Systematically Test and Evaluate Your LLMs Apps // MLOps Podcast #269 with Gideon Mendels, CEO of Comet. // Abstract When building LLM Applications, Developers need to take a hybrid approach from both ML and SW Engineering best practices. They need to define eval metrics and track their entire experimentation to see what is and is not working. They also need to define comprehensive unit tests for their particular use-case so they can confidently check if their LLM App is ready to be deployed. // Bio Gideon Mendels is the CEO and co-founder of Comet, the leading solution for managing machine learning workflows from experimentation to production. He is a computer scientist, ML researcher and entrepreneur at his core. Before Comet, Gideon co-founded GroupWize, where they trained and deployed NLP models processing billions of chats. His journey with NLP and Speech Recognition models began at Columbia University and Google where he worked on hate speech and deception detection. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.comet.com/site/ All the Hard Stuff with LLMs in Product Development // Phillip Carter // MLOps Podcast #170: https://youtu.be/DZgXln3v85s
Opik by Comet: https://www.comet.com/site/products/opik/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Gideon on LinkedIn: https://www.linkedin.com/in/gideon-mendels/ Timestamps: [00:00] Gideon's preferred coffee [00:17] Takeaways [01:50] A huge shout-out to Comet ML for sponsoring this episode! [02:09] Please like, share, leave a review, and subscribe to our MLOps channels! [03:30] Evaluation metrics in AI [06:55] LLM Evaluation in Practice [10:57] LLM testing methodologies [16:56] LLM as a judge [18:53] OPIC track function overview [20:33] Tracking user response value [26:32] Exploring AI metrics integration [29:05] Experiment tracking and LLMs [34:27] Micro Macro collaboration in AI [38:20] RAG Pipeline Reproducibility Snapshot [40:15] Collaborative experiment tracking [45:29] Feature flags in CI/CD [48:55] Labeling challenges and solutions [54:31] LLM output quality alerts [56:32] Anomaly detection in model outputs [1:01:07] Wrap up
Raj Rikhy is a Senior Product Manager at Microsoft AI + R, enabling deep reinforcement learning use cases for autonomous systems. Previously, Raj was the Group Technical Product Manager in the CDO for Data Science and Deep Learning at IBM. Prior to joining IBM, Raj has been working in product management for several years - at Bitnami, Appdirect and Salesforce. // MLOps Podcast #268 with Raj Rikhy, Principal Product Manager at Microsoft. // Abstract In this MLOps Community podcast, Demetrios chats with Raj Rikhy, Principal Product Manager at Microsoft, about deploying AI agents in production. They discuss starting with simple tools, setting clear success criteria, and deploying agents in controlled environments for better scaling. Raj highlights real-time uses like fraud detection and optimizing inference costs with LLMs, while stressing human oversight during early deployment to manage LLM randomness. The episode offers practical advice on deploying AI agents thoughtfully and efficiently, avoiding over-engineering, and integrating AI into everyday applications. // Bio Raj is a Senior Product Manager at Microsoft AI + R, enabling deep reinforcement learning use cases for autonomous systems. Previously, Raj was the Group Technical Product Manager in the CDO for Data Science and Deep Learning at IBM. Prior to joining IBM, Raj has been working in product management for several years - at Bitnami, Appdirect and Salesforce. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.microsoft.com/en-us/research/focus-area/ai-and-microsoft-research/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Raj on LinkedIn: https://www.linkedin.com/in/rajrikhy/
//Abstract If there is one thing that is true, it is data is constantly changing. How can we keep up with these changes? How can we make sure that every stakeholder has visibility? How can we create a culture of understanding around data change management? //Bio - Benjamin Rogojan: Data Science And Engineering Consultant @ Seattle Data Guy - Chad Sanderson: CEO & Co-Founder @ Gable - Christophe Blefari: CTO & Co-founder @ NAO - Maggie Hays: Founding Community Product Manager, DataHub @ Acryl Data A big thank you to our Premium Sponsors @Databricks , @tecton8241 , & @onehouseHQ for their generous support!
The AI Dream Team: Strategies for ML Recruitment and Growth // MLOps Podcast #267 with Jelmer Borst, Analytics & Machine Learning Domain Lead, and Daniela Solis, Machine Learning Product Owner, of Picnic. // Abstract Like many companies, Picnic started out with a small, central data science team. As this grows larger, focussing on more complex models, it questions the skillsets & organisational set up. Use an ML platform, or build ourselves? A central team vs. embedded? Hire data scientists vs. ML engineers vs. MLOps engineers How to foster a team culture of end-to-end ownership How to balance short-term & long-term impact // Bio Jelmer Borst Jelmer leads the analytics & machine learning teams at Picnic, an app-only online groceries company based in The Netherlands. Whilst his background is in aerospace engineering, he was looking for something faster-paced and found that at Picnic. He loves the intersection of solving business challenges using technology & data. In his free time loves to cook food and tinker with the latest AI developments. Daniela Solis Morales As a Machine Learning Lead at Picnic, I am responsible for ensuring the success of end-to-end Machine Learning systems. My work involves bringing models into production across various domains, including Personalization, Fraud Detection, and Natural Language Processing. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jelmer on LinkedIn: https://www.linkedin.com/in/japborst Connect with Daniela on LinkedIn: https://www.linkedin.com/in/daniela-solis-morales/
Francisco Ingham, LLM consultant, NLP developer, and founder of Pampa Labs. Making Your Company LLM-native // MLOps Podcast #266 with Francisco Ingham, Founder of Pampa Labs. // Abstract Being an LLM-native is becoming one of the key differentiators among companies, in vastly different verticals. Everyone wants to use LLMs, and everyone wants to be on top of the current tech but - what does it really mean to be LLM-native? LLM-native involves two ends of a spectrum. On the one hand, we have the product or service that the company offers, which surely offers many automation opportunities. LLMs can be applied strategically to scale at a lower cost and offer a better experience for users. But being LLM-native not only involves the company's customers, it also involves each stakeholder involved in the company's operations. How can employees integrate LLMs into their daily workflows? How can we as developers leverage the advancements in the field not only as builders but as adopters? We will tackle these and other key questions for anyone looking to capitalize on the LLM wave, prioritizing real results over the hype. // Bio Currently working at Pampa Labs, where we help companies become AI-native and build AI-native products. Our expertise lies on the LLM-science side, or how to build a successful data flywheel to leverage user interactions to continuously improve the product. We also spearhead, pampa-friends - the first Spanish-speaking community of AI Engineers. Previously worked in management consulting, was a TA in fastai in SF, and led the cross-AI + dev tools team at Mercado Libre. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: pampa.ai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Francisco on LinkedIn: https://www.linkedin.com/in/fpingham/ Timestamps: [00:00] Francisco's preferred coffee [00:13] Takeaways [00:37] Please like, share, leave a review, and subscribe to our MLOps channels! [00:51] A Literature Geek [02:41] LLM-native company [03:54] Integrating LLM in workflows [07:21] Unexpected LLM applications [10:38] LLM's in development process [14:00] Vibe check to evaluation [15:36] Experiment tracking optimizations [20:22] LLMs as judges discussion [24:43] Presentaciones automatizadas para podcast [27:48] AI operating system and agents [31:29] Importance of SEO expertise [35:33] Experimentation and evaluation [39:20] AI integration strategies [41:50] RAG approach spectrum analysis [44:40] Search vs Retrieval in AI [49:02] Recommender Systems vs RAG [52:08] LLMs in recommender systems [53:10] LLM interface design insights
Simba Khadder is the Founder & CEO of Featureform. He started his ML career in recommender systems where he architected a multi-modal personalization engine that powered 100s of millions of user’s experiences. Unpacking 3 Types of Feature Stores // MLOps Podcast #265 with Simba Khadder, Founder & CEO of Featureform. // Abstract Simba dives into how feature stores have evolved and how they now intersect with vector stores, especially in the world of machine learning and LLMs. He breaks down what embeddings are, how they power recommender systems, and why personalization is key to improving LLM prompts. Simba also sheds light on the difference between feature and vector stores, explaining how each plays its part in making ML workflows smoother. Plus, we get into the latest challenges and cool innovations happening in MLOps. // Bio Simba Khadder is the Founder & CEO of Featureform. After leaving Google, Simba founded his first company, TritonML. His startup grew quickly and Simba and his team built ML infrastructure that handled over 100M monthly active users. He instilled his learnings into Featureform’s virtual feature store. Featureform turns your existing infrastructure into a Feature Store. He’s also an avid surfer, a mixed martial artist, a published astrophysicist for his work on finding Planet 9, and he ran the SF marathon in basketball shoes. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: featureform.comBigQuery Feature Store // Nicolas Mauti // MLOps Podcast #255: https://www.youtube.com/watch?v=NtDKbGyRHXQ&ab_channel=MLOps.community --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Simba on LinkedIn: https://www.linkedin.com/in/simba-k/ Timestamps: [00:00] Simba's preferred coffee [00:08] Takeaways [02:01] Coining the term 'Embedding' [07:10] Dual Tower Recommender System [10:06] Complexity vs Reliability in AI [12:39] Vector Stores and Feature Stores [17:56] Value of Data Scientists [20:27] Scalability vs Quick Solutions [23:07] MLOps vs LLMOps Debate [24:12] Feature Stores' current landscape [32:02] ML lifecycle challenges and tools [36:16] Feature Stores bundling impact [42:13] Feature Stores and BigQuery [47:42] Virtual vs Literal Feature Store [50:13] Hadoop Community Challenges [52:46] LLM data lifecycle challenges [56:30] Personalization in prompting usage [59:09] Contextualizing company variables [1:03:10] DSPy framework adoption insights [1:05:25] Wrap up
Stefano Bosisio is an accomplished MLOps Engineer with a solid background in Biomedical Engineering, focusing on cellular biology, genetics, and molecular simulations. Reinvent Yourself and Be Curious // MLOps Podcast #264 with Stefano Bosisio, MLOps Engineer at Synthesia. // Abstract This talk goes through Stefano's experience, to be an inspirational source for whoever wants to jump on a career in the MLOps sector. Moreover, Stefano will also introduce his MLOps Course on the MLOps community platform. // Bio Sai Bharath Gottam Stefano Bosisio is an MLOps Engineer, with a versatile background that ranges from biomedical engineering to computational chemistry and data science. Stefano got an MSc in biomedical engineering from the Polytechnic of Milan, focusing on cellular biology, genetics, and molecular simulations. Then, he landed in Scotland, in Edinburgh, to earn a PhD in chemistry from the University of Edinburgh, where he developed robust physical theories and simulation methods, to understand and unlock the drug discovery problem. After completing his PhD, Stefano transitioned into Data Science, where he began his career as a data scientist. His interest in machine learning engineering grew, leading him to specialize in building ML platforms that drive business success. Stefano's expertise bridges the gap between complex scientific research and practical machine learning applications, making him a key figure in the MLOps field. Bonus points beyond data: Stefano, as a proper Italian, loves cooking and (mainly) baking, playing the piano, crocheting and running half-marathons. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://medium.com/@stefanobosisio1First MLOps Stack Course: https://learn.mlops.community/courses/languages/your-first-mlops-stack/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stefano on LinkedIn: https://www.linkedin.com/in/stefano-bosisio1/ Timestamps: [00:00] Stephano's preferred coffee [00:12] Takeaways [01:06] Stephano's MLOps Course [01:47] From Academia to AI Industry [09:10] Data science and platforms [16:53] Persistent MLOps challenges [21:23] Internal evangelization for success [24:21] Adapt communication skills to diverse individual needs
[29:43] Key components of ML pipelines are essentia
l[33:47] Create a generalizable AI training pipeline with Kubeflow
[35:44] Consider cost-effective algorithms and deployment methods
[39:02] Agree with dream platform; LLMs require simple microservice
[42:48] Auto scaling: crucial, tricky, prone to issues
[46:28] Auto-scaling issues with Apache Beam data pipelines
[49:49] Guiding students through MLOps with practical experience
[53:16] Bulletproof Problem Solving: Decision trees for problem analysis
[55:03] Evaluate tools critically; appreciate educational opportunities
[57:01] Wrap up
Global Feature Store: Optimizing Locally and Scaling Globally at Delivery Hero // MLOps Podcast #263 with Delivery Hero's Gottam Sai Bharath, Senior Machine Learning Engineer & Cole Bailey, ML Platform Engineering Manager. // Abstract Delivery Hero innovates locally within each department to develop MLOps practices most effective in that particular context. We also discuss our efforts to reduce redundancy and inefficiency across the company. Hear about our experiences in creating multiple micro feature stores within our departments, and our goal to unify these into a Global Feature Store that is more powerful when combined. // Bio Sai Bharath Gottam With a passion for translating complex technical concepts into practical solutions, Sai excels at making intricate topics accessible and engaging. As a Senior Machine Learning Engineer at Delivery Hero, Sai works on cutting-edge machine learning platforms that guarantee seamless delivery experiences. Always eager to share insights and innovations, Sai is committed to making technology understandable and enjoyable for all. Cole Bailey Bridging data science and production-grade software engineering. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.deliveryhero.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sai on LinkedIn: https://www.linkedin.com/in/sai-bharath-gottam/ Connect with Cole on LinkedIn: www.linkedin.com/in/cole-bailey Timestamps: [00:00] Sai and Cole's preferred coffee [00:42] Takeaways [01:51] Please like, share, leave a review, and subscribe to our MLOps channels! [02:08] Life changes in Delivery Hero [05:21] Global Feature Store and Pandora [12:21] Tech integration strategies [20:08] Defining Feature and Feature Store [22:46] Feature Store vs Data Platform [26:26] Features are discoverable [32:56] Onboarding and Feature Testing [36:00] Data consistency [41:07] Future Vision Feature Store [44:17] Multi-cloud strategies [46:33] Wrap up
Adam Kamor is the Co-founder of Tonic, a company that specializes in creating mock data that preserves secure datasets.
RAG Quality Starts with Data Quality // MLOps Podcast #262 with Adam Kamor, Co-Founder & Head of Engineering of Tonic.ai. // Abstract Dive into what makes Retrieval-Augmented Generation (RAG) systems tick—and it all starts with the data. We’ll be talking with an expert in the field who knows exactly how to transform messy, unstructured enterprise data into high-quality fuel for RAG systems. Expect to learn the essentials of data prep, uncover the common challenges that can derail even the best-laid plans, and discover some insider tips on how to boost your RAG system’s performance. We’ll also touch on the critical aspects of data privacy and governance, ensuring your data stays secure while maximizing its utility. If you’re aiming to get the most out of your RAG systems or just curious about the behind-the-scenes work that makes them effective, this episode is packed with insights that can help you level up your game. // Bio Adam Kamor, PhD, is the Co-founder and Head of Engineering of Tonic.ai. Since completing his PhD in Physics at Georgia Tech, Adam has committed himself to enabling the work of others through the programs he develops. In his roles at Microsoft and Kabbage, he handled UI design and led the development of new features to anticipate customer needs. At Tableau, he played a role in developing the platform’s analytics/calculation capabilities. As a founder of Tonic.ai, he is leading the development of unstructured data solutions that are transforming the work of fellow developers, analysts, and data engineers alike. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.tonic.ai Various topics about RAG and LLM security are available on Tonic.ai's blogs: https://www.tonic.ai/blog https://www.tonic.ai/blog/how-to-prevent-data-leakage-in-your-ai-applications-with-tonic-textual-and-snowpark-container-services https://www.tonic.ai/blog/rag-evaluation-series-validating-the-rag-performance-of-the-openais-rag-assistant-vs-googles-vertex-search-and-conversation https://www.youtube.com/watch?v=5xdyt4oRONU https://www.tonic.ai/blog/what-is-retrieval-augmented-generation-the-benefits-of-implementing-rag-in-using-llms --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Adam on LinkedIn: https://www.linkedin.com/in/adam-kamor-85720b48/ Timestamps: [00:00] Adam's preferred coffee [00:24] Takeaways [00:59] Huge shout out to Tonic.ai for supporting the community! [01:03] Please like, share, leave a review, and subscribe to our MLOps channels! [01:18] Naming a product [03:38] Tonic Textual [08:00] Managing PII and Data Safety [10:16] Chunking strategies for context [14:19] Data prep for RAG [17:20] Data quality in AI systems [20:58] Data integrity in PDFs [27:12] Ensuring chatbot data freshness [33:02] Managed PostgreSQL and Vector DB [34:49] RBAC database vs file access [37:35] Slack AI data leakage solutions [42:26] Hot swapping [46:06] LLM security concerns [47:03] Privacy management best practices [49:02] Chatbot design patterns [50:39] RAG growth and impact [52:40] Retrieval Evaluation best practices [59:20] Wrap up
Jonathan Rioux is a Managing Principal of AI Consulting for EPAM Systems, where he advises clients on how to get from idea to realized AI products with the minimum of fuss and friction. Who's MLOps for Anyway? // MLOps Podcast #261 with Jonathan Rioux, Managing Principal, AI Consulting at EPAM Systems. // Abstract The year is 2024 and we are all staring into the cliff towards the abyss of disillusionment for Generative AI. Every organization, developer, and AI-adjacent individual is now talking about "making AI real" and "turning a ROI on AI initiatives". MLOps and LLMOps are taking the stage as the solution; equip your AI teams with the best tools money can buy, grab tokens by the fistful, and look at value raking in. Sounds familiar and eerily similar to the previous ML hype cycles? From solo devs to large organizations, how can we avoid the same pitfalls as last time and get out of the endless hamster wheel? // Bio Jonathan is a Managing Principal of AI Consulting for EPAM, where he advises client on how to get from idea to realized AI products with the minimum of fuss and friction. He's obsessed with the mental models of ML and how to organize harmonious AI practices. Jonathan published "Data Analysis with Python and PySpark" (Manning, 2022). // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: raiks.ca --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jonathan on LinkedIn: https://www.linkedin.com/in/jonathanrx/ Timestamps: [00:00] Jonathan's preferred coffee [00:25] Takeaways [01:44] MLOps as not being sexy [03:49] Do not conflate MLOps with ROI [06:21] ML Certification Business Idea [11:02] AI Adoption Missteps [15:40] Slack AI Privacy Risks [18:17] Decentralized AI success [22:00] Michelangelo Hub-Spoke Model [27:45] Engineering tools for everyone [33:38 - 35:20] SAS Ad [35:21] POC to ROI transition [42:08] Repurposing project learnings [46:24] Balancing Innovation and ROI [55:35] Using classification model [1:00:24] Chatbot evolution comparison [1:01:20] Balancing Automation and Trust [1:06:30] Manual to AI transition [1:09:57] Wrap up
Shiva Bhattacharjee is the Co-founder and CTO of TrueLaw, where we are building bespoke models for law firms for a wide variety of tasks. Alignment is Real // MLOps Podcast #260 with Shiva Bhattacharjee, CTO of TrueLaw Inc. // Abstract If the off-the-shelf model can understand and solve a domain-specific task well enough, either your task isn't that nuanced or you have achieved AGI. We discuss when is fine-tuning necessary over prompting and how we have created a loop of sampling - collecting feedback - fine-tuning to create models that seem to perform exceedingly well in domain-specific tasks. // Bio 20 years of experience in distributed and data-intensive systems spanning work at Apple, Arista Networks, Databricks, and Confluent. Currently CTO at TrueLaw where we provide a framework to fold in user feedback, such as lawyer critiques of a given task, and fold them into proprietary LLM models through fine-tuning mechanics, resulting in 7-10x improvements over the base model. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.truelaw.ai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Shiva on LinkedIn: https://www.linkedin.com/in/shivabhattacharjee/ Timestamps: [00:00] Shiva's preferred coffee [00:58] Takeaways [01:17] DSPy Implementation [04:57] Evaluating DSPy risks [08:13] Community-driven DSPy tool [12:19] RAG implementation strategies [17:02] Cost-effective embedding fine-tuning [18:51] AI infrastructure decision-making [24:13] Prompt data flow evolution [26:32] Buy vs build decision [30:45] Tech stack insights [38:20] Wrap up
Vikram Rangnekar is an open-source software developer focused on simplifying LLM integration. He created LLMClient, a TypeScript library inspired by Stanford's DSP paper. With years of experience building complex LLM workflows, he previously worked as a senior software engineer at LinkedIn on Ad Serving. Ax a New Way to Build Complex Workflows with LLMs // MLOps Podcast #259 with Vikram Rangnekar, Software Engineer at Stealth. // Abstract Ax is a new way to build complex workflows with LLMs. It's a typescript library based on research done in the Stanford DSP paper. Concepts such as prompt signatures, prompt tuning, and composable prompts help you build RAG and agent-powered ideas that have till now been hard to build and maintain. Ax is designed for production usage. // Bio Vikram builds open-source software. Currently working on making it easy to build with LLMs. Created Ax a typescript library that abstracts over all the complexity of LLMs, it is based on the research done in the Stanford DSP paper. Worked extensively with LLMs over the last few years to build complex workflows. Previously worked as a senior software engineer with LinkedIn on Ad Serving. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links The unofficial DSPy framework. Build LLM-powered Agents and "Agentic workflows" based on the Stanford DSP paper: https://axllm.dev All the Hard Stuff with LLMs in Product Development // Phillip Carter // MLOps Podcast #170: https://youtu.be/DZgXln3v85s --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vikram on LinkedIn: https://www.linkedin.com/in/vikramr Timestamps: [00:00] Vikram preferred coffee [00:41] Takeaways [01:05] Data Engineering for AI/ML Conference Ad [01:41] Vikram's work these days [04:54] Fine-tuned Model insights [06:22] Java Script tool evolution [16:14] DSP knowledge distillation [17:34] DSP vs Manual examples [22:53] Optimizing task context [27:58] API type validation explained [30:25] LLM value and innovation [34:22] Navigating complex systems [37:30] DSP code generators explained [40:56] Exploring LLM personas [42:45] Optimizing small agents [43:32] Complex task assistance [49:53] Wrap up
MLOps Coffee Sessions #177 with Mohamed Abusaid and Mara Pometti, Building in Production Human-centred GenAI Solutions sponsored by QuantumBlack, AI by McKinsey. // Abstract Trust is paramount in the adoption of new technologies, especially in the realm of education. Mohamed and Mara shed light on the importance of AI governance programs and establishing AI governance boards to ensure safe and ethical use of technology while managing associated risks. They discuss the impact on customers, potential risks, and mitigation strategies that organizations must consider to protect their brand reputation and comply with regulations. // Bio Mara Pometti Mara is an Associate Design Director at McKinsey & Company, where she helps organisations drive AI adoption through human-centered methods. She defines herself as a data-savvy humanist. Her practice spans across AI, data journalism, and design with the overarching objective of finding the strategic intersection between AI models and human intents to implement responsible AI systems that move organisations forward. Previously, she led the AI Strategy practice at IBM, where she also developed the company’s first-ever data storytelling program. Yet, by background, she is a data journalist. She worked as a data journalist for agencies and newsrooms like Aljazeera. Mara lectured at many universities about how to humanize AI, including the London School of Economics. Her books and writing explore how to weave a humanistic approach to AI development. Mohamed Abusaid Am Mohamed, a tech enthusiast, hacker, avid traveler, and foodie all rolled into one individual. Built his first website when he was 9 and fell in love with computers and the internet ever since. Graduated with computer science from university although dabbled in electrical, electronic, and network engineering before that. When he's not reading up on the latest tech conversations and products on Hacker News, Mohamed spends his time traveling to new destinations and exploring their cuisine and culture. Mohamed works with different companies helping them tackle challenges in developing, deploying, and scaling their analytics to reach its potential. Some topics he's enthusiastic about include MLOps, DataOps, GenerativeAI, Product thinking, and building cross-functional teams to deliver user-first products. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links QuantumBlack, AI by McKinsey: https://www.mckinsey.com/capabilities/quantumblack/how-we-help-clients --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stephen-batifol/ Connect with Mara on LinkedIn: https://www.linkedin.com/in/mara-pometti Connect with Mohamed on LinkedIn: https://www.linkedin.com/in/mabusaid/
Markus Stoll is the Co-Founder of Renumics and the developer behind the open-source interactive ML dataset exploration tool, Spotlight. He shares insights on:
AI in Engineering and Manufacturing
Interactive ML Data Visualization
ML Data Exploration
Follow Markus for hands-on articles about leveraging ML while keeping a strong focus on data.
Visualize - Bringing Structure to Unstructured Data // MLOps Podcast #258 with Markus Stoll, CTO of Renumics. A huge thank you to SAS for their generous support! // Abstract This talk is about how data visualization and embeddings can support you in understanding your machine-learning data. We explore methods to structure and visualize unstructured data like text, images, and audio for applications ranging from classification and detection to Retrieval-Augmented Generation. By using tools and techniques like UMAP to reduce data dimensions and visualization tools like Renumics Spotlight, we aim to make data analysis for ML easier. Whether you're dealing with interpretable features, metadata, or embeddings, we'll show you how to use them all together to uncover hidden patterns in multimodal data, evaluate the model performance for data subgroups, and find failure modes of your ML models. // Bio Markus Stoll began his career in the industry at Siemens Healthineers, developing software for the Heavy Ion Therapy Center in Heidelberg. He learned about software quality while developing a treatment machine weighing over 600 tons. He earned a Ph.D., focusing on combining biomechanical models with statistical models, through which he learned how challenging it is to bridge the gap between research and practical application in the healthcare domain. Since co-founding Renumics, he has been active in the field of AI for Engineering, e.g., AI for Computer Aided Engineering (CAE), implementing projects, contributing to their open-source library for data exploration for ML datasets (Renumics Spotlight) and writing articles about data visualization. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://renumics.com/ MLSecOps Community: https://community.mlsecops.com/ Blogs: https://towardsdatascience.com/visualize-your-rag-data-evaluate-your-retrieval-augmented-generation-system-with-ragas-fc2486308557 : https://medium.com/itnext/how-to-explore-and-visualize-ml-data-for-object-detection-in-images-88e074f46361 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Markus on LinkedIn: https://www.linkedin.com/in/markus-stoll-b39a42138/
MLOps for GenAI Applications // Special MLOps Podcast episode with Demetrios Brinkmann, Chief Happiness Engineer at MLOps Community. // Abstract Demetrios explores common themes in ML model testing with insights from Erica Greene (Yahoo News), Matar Haller (ActiveFence), Mohamed Elgendy (Kolena), and Catherine Nelson (Freelance Data Scientist). They discuss tiered test cases, functional testing for hate speech, differences between AI and traditional software testing, and the complexities of evaluating LLMs. Demetrios wraps up by inviting feedback and promoting an upcoming virtual conference on data engineering for AI and ML. // Bio At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps Community Podcasts. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Balancing Speed and Safety // Panel // AIQCON - https://youtu.be/c81puRgu3Kw AI For Good - Detecting Harmful Content at Scale // Matar Haller // MLOps Podcast #246 - https://youtu.be/wLKlZ6yHg1k What is AI Quality? // Mohamed Elgendy // MLOps Podcast #229 - https://youtu.be/-Jdmq4DiOew All Data Scientists Should Learn Software Engineering Principles // Catherine Nelson // Podcast #245 - https://youtu.be/yP6Eyny7p20 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Timestamps: [00:00] Exploring common themes in MLOps community [00:49] Common patterns about model output and testing [01:34] Tiered test case strategy [03:05] Functional testing for models [05:24] Testing coverage and quality [07:47] Evaluating LLMs challenges [08:35] Please like, share, leave a review, and subscribe to our MLOps channels!
Sean Morgan is an active open-source contributor and maintainer and is the special interest group lead for TensorFlow Addons. Learn more about the platform for end-to-end AI Security at https://protectai.com/. MLSecOps is Fundamental to Robust AI Security Posture Management (AISPM) // MLOps Podcast #257 with Sean Morgan, Chief Architect at Protect AI. // Abstract MLSecOps, which is the practice of integrating security practices into the AIML lifecycle (think infusing MLOps with DevSecOps practices), is a critical part of any team’s AI Security Posture Management. In this talk, we’ll discuss how to threat model realistic AIML security risks, how you can measure your organization’s AI Security Posture, and most importantly how you can improve that security posture through the use of MLSecOps. // Bio Sean Morgan is the Chief Architect at Protect AI. In prior roles he's led production AIML deployments in the semiconductor industry, evaluated adversarial machine learning defenses for DARPA research programs, and most recently scaled customers on interactive machine learning solutions at AWS. In his free time, Sean is an active open-source contributor and maintainer, and is the special interest group lead for TensorFlow Addons. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Sean's GitHub: https://github.com/seanpmorgan MLSecOps Community: https://community.mlsecops.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sean on LinkedIn: https://www.linkedin.com/in/seanmorgan/ Timestamps: [00:00] Sean's preferred coffee [00:10] Takeaways [01:39] Register for the Data Engineering for AI/ML Conference now! [02:21] KubeCon Paris: Emphasis on security and AI [05:00] Concern about malicious data during training process [09:29] Model builders, security, pulling foundational models, nuances [12:13] Hugging Face research on security issues [15:00] Inference servers exposed; potential for attack [19:45] Balancing ML and security processes for ease [23:23] Model artifact security in enterprise machine learning [25:04] Scanning models and datasets for vulnerabilities [29:23] Ray's user interface vulnerabilities lead to attacks [32:07] ML Flow vulnerabilities present significant server risks [36:04] Data ops essential for machine learning security [37:32] Prioritized security in model and data deployment [40:46] Automated scanning tool for improved antivirus protection [42:00] Wrap up
Harcharan Kabbay is a Data Scientist & AI/ML Engineer with Expertise in MLOps, Kubernetes, and DevOps, Driving End-to-End Automation and Transforming Data into Actionable Insights.
MLOps for GenAI Applications // MLOps Podcast #256 with Harcharan Kabbay, Lead Machine Learning Engineer at World Wide Technology. // Abstract The discussion begins with a brief overview of the Retrieval-Augmented Generation (RAG) framework, highlighting its significance in enhancing AI capabilities by combining retrieval mechanisms with generative models. The podcast further explores the integration of MLOps, focusing on best practices for embedding the RAG framework into a CI/CD pipeline. This includes ensuring robust monitoring, effective version control, and automated deployment processes that maintain the agility and efficiency of AI applications. A significant portion of the conversation is dedicated to the importance of automation in platform provisioning, emphasizing tools like Terraform. The discussion extends to application design, covering essential elements such as key vaults, configurations, and strategies for seamless promotion across different environments (development, testing, and production). We'll also address how to enhance the security posture of applications through network firewalls, key rotation, and other measures. Let's talk about the power of Kubernetes and related tools to aid a good application design. The podcast highlights the principles of good application design, including proper observability and eliminating single points of failure. I would share strategies to reduce development time by creating templates for GitHub repositories by application types to be re-used, also templates for pull requests, thereby minimizing human errors and streamlining the development process. // Bio Harcharan is an AI and machine learning expert with a robust background in Kubernetes, DevOps, and automation. He specializes in MLOps, facilitating the adoption of industry best practices and platform provisioning automation. With extensive experience in developing and optimizing ML and data engineering pipelines, Harcharan excels at integrating RAG-based applications into production environments. His expertise in building scalable, automated AI systems has empowered the organization to enhance decision-making and problem-solving capabilities through advanced machine-learning techniques. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Harcharan's Medium - https://medium.com/@harcharan-kabbay Data Engineering for AI/ML Conference: https://home.mlops.community/home/events/dataengforai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Harcharan on LinkedIn: https://www.linkedin.com/in/harcharankabbay/locale=en_US
Nicolas Mauti is an MLOps Engineer from Lyon (France), Working at Malt. BigQuery Feature Store // MLOps Podcast #255 with Nicolas Mauti, Lead MLOps at Malt. // Abstract Need a feature store for your AI/ML applications but overwhelmed by the multitude of options? Think again. In this talk, Nicolas shares how they solved this issue at Malt by leveraging the tools they already had in place. From ingestion to training, Nicolas provides insights on how to transform BigQuery into an effective feature management system. We cover how Nicolas' team designed their feature tables and addressed challenges such as monitoring, alerting, data quality, point-in-time lookups, and backfilling. If you’re looking for a simpler way to manage your features without the overhead of additional software, this talk is for you. Discover how BigQuery can handle it all! // Bio Nicolas Mauti is the go-to guy for all things related to MLOps at Malt. With a knack for turning complex problems into streamlined solutions and over a decade of experience in code, data, and ops, he is a driving force in developing and deploying machine learning models that actually work in production. When he's not busy optimizing AI workflows, you can find him sharing his knowledge at the university. Whether it's cracking a tough data challenge or cracking a joke, Nicolas knows how to keep things interesting. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Nicolas' Medium - https://medium.com/@nmauti Data Engineering for AI/ML Conference: https://home.mlops.community/home/events/dataengforai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Nicolas on LinkedIn: https://www.linkedin.com/in/nicolasmauti/?locale=en_US Timestamps: [00:00] Nicolas' preferred beverage [00:35] Takeaways [02:25] Please like, share, leave a review, and subscribe to our MLOps channels! [02:57] BigQuery end goal [05:00] BigQuery pain points [10:14] BigQuery vs Feature Stores [12:54] Freelancing Rate Matching issues [16:43] Post-implementation pain points [19:39] Feature Request Process [20:45] Feature Naming Consistency [23:42] Feature Usage Analysis [26:59] Anomaly detection in data [28:25] Continuous Model Retraining Process [30:26] Model misbehavior detection [33:01] Handling model latency issues [36:28] Accuracy vs The Business [38:59] BigQuery cist-benefit analysis [42:06] Feature stores cost savings [44:09] When not to use BigQuery [46:20] Real-time vs Batch Processing [49:11] Register for the Data Engineering for AI/ML Conference now! [50:14] Wrap up
Design and Development Principles for LLMOps // MLOps Podcast #254 with Andy McMahon, Director - Principal AI Engineer at Barclays Bank. A huge thank you to SAS for their generous support! // Abstract As we move from MLOps to LLMOps we need to double down on some fundamental software engineering practices, as well as augment and add to these with some new techniques. In this case, let's talk about this! // Bio Andy is a Principal AI Engineer, working in the new AI Center of Excellence at Barclays Bank. Previously he was Head of MLOps for NatWest Group, where he led their MLOps Centre of Excellence and helped build out their MLOps platform and processes across the bank. Andy is also the author of Machine Learning Engineering with Python, a hands-on technical book published by Packt. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Andy's book - https://packt.link/w3JKL Andy's Medium - https://medium.com/@andrewpmcmahon629 SAS: https://www.sas.com/en_us/home.html SAS® Decision Builder: https://www.sas.com/en_us/offers/23q4/microsoft-fabric.html Data Engineering for AI/ML Conference: https://home.mlops.community/home/events/dataengforai Harnessing MLOps in Finance // Michelle Marie Conway // MLOps Podcast Coffee #174: https://youtu.be/nIEld_Q6L-0The Lean Startup: How Today's Entrepreneurs Use Continuous Innovation to Create Radically Successful Businesses book by Eric Ries: https://www.amazon.co.jp/-/en/Eric-Ries/dp/0307887898 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Andy on LinkedIn: https://www.linkedin.com/in/andrew-p-mcmahon/
Timestamps: [00:00] Andy's preferred coffee [00:09] Takeaways [02:04] Andy's book as an Oxford curriculum [06:13] Register for the Data Engineering for AI/ML Conference now! [07:04] The Life Cycle of AI Executives Course [09:55] MLOps as a term [11:53] Tooling vs Process Culture [15:01] Open source benefits [17:15] End goal flexibility [20:06] Hybrid Cloud Strategy Overview [21:11] ROI for tool upgrades [25:41] Long-term projects comparison [29:02 - 30:48] SAS Ad [30:49] AI and ML Integration [35:40] Hybrid AI Integration Insights [42:18] Tech trends vs Practicality [44:39] Gen AI Tooling Debate [51:57] Vanity metrics overview [55:22] Tech business alignment strategy [58:45] Aligning teams for ROI [1:01:35] Communication mission effectively [1:03:45] Enablement metrics [1:06:38] Prioritizing use cases [1:09:47] Wrap up
// Abstract Data is the foundation of AI. To ensure AI performs as expected, high-quality data is essential. In this panel discussion, Chad, Maria, Joe, and Pushkar hosted by Sam Partee will explore strategies for obtaining and maintaining high-quality data, as well as common pitfalls to avoid when using data for AI models. // Panelists - Samuel Partee: Principal Applied AI Engineer @ Redis - Chad Sanderson: CEO & Co-Founder @ Gable - Joe Reis: CEO/Co-Founder @ Ternary Data - Maria Zhang: CEO Cofounder @ Proactive AI Lab Inc - Pushkar Garg: Staff Machine Learning Engineer @ Clari Inc.
Yuri Plotkin is a Biomedical Engineer and Machine Learning Scientist and the author of The Variational Book.
The Variational Book // MLOps Podcast #253 with Yuri Plotkin, an ML Scientist. // Abstract Curiosity has been the underlying thread in Yuri's life and interests. With the explosion of Generative AI, Yuri was fascinated by the topic and decided he needed to learn more. Yuri pursued learning by reading, deriving, and understanding seminal papers within the last generation. The endeavors culminated in the writing of a book on the topic, The Variational Book, which Yuri expects to release shortly in the coming months. A bit of detail about the topics he covers can be found here: www.thevariationalbook.com. // Bio Evolved from biomedical engineer to wet-lab scientist, and more recently transitioned Yuri's career to computer science with the last 10+ years developing projects at the intersection of medicine, life sciences and machine learning. Yuri's educational background is in Biomedical Engineering, at Columbia University (M.S.) and University of California, San Diego (B.S.). Current interests include generative AI, diffusion models, and LLMs. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://plotkiny.github.io/ The Variational Book: www.thevariationalbook.com SAS: https://www.sas.com/en_us/home.html SAS® Decision Builder: https://www.sas.com/en_us/offers/23q4/microsoft-fabric.html
Data Engineering for AI/ML Conference: https://home.mlops.community/home/events/dataengforai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Yuri on LinkedIn: http://www.linkedin.com/in/yuri-plotkin/
// Abstract Attracting and retaining top AI talent is essential for staying competitive. This panel will explore crafting and communicating a compelling vision that aligns with the organization's evolving needs, inspiring potential hires and motivating current employees. The discussion will offer actionable strategies for sourcing top talent, adapting to changing needs, and maintaining company alignment. Attendees will learn best practices for attracting AI professionals, creating an attractive employer brand, and enhancing talent acquisition and retention strategies. Lastly, the panel will cover structuring and organizing the AI team as it grows to ensure alignment with business goals. This includes optimal team configurations, leadership roles, and processes that support collaboration and innovation, enabling sustained growth and success. // PANELISTS Ashley Antonides: Associate Research Director, AI/ML @ Two Six Technologies Olga Beregovaya: VP, AI @ Smartling Shailvi Wakhlu: Founder @ Shailvi Ventures LLC A big thank you to our Premium Sponsors Google Cloud & Databricks for their generous support!
Ron Heichmn is an AI researcher specializing in generative AI, AI alignment, and prompt engineering. At SentinelOne, Ron actively monitors emerging research to identify and address potential vulnerabilities in our AI systems, focusing on unsupervised and scalable evaluations to ensure robustness and reliability.
Harnessing AI APIs for Safer, Accurate, & Reliable Applications // MLOps Podcast #252 with Ron Heichman, Machine Learning Engineer at SentinelOne. // Abstract Integrating AI APIs effectively is pivotal for building applications that leverage LLMs, especially given the inherent issues with accuracy, reliability, and safety that LLMs often exhibit. I aim to share practical strategies and experiences for using AI APIs in production settings, detailing how to adapt these APIs to specific use cases, mitigate potential risks, and enhance performance. The focus will be testing, measuring, and improving quality for RAG or knowledge workers utilizing AI APIs. // Bio Ron Heichman is an AI researcher and engineer dedicated to advancing the field through his work on prompt injection at Preamble, where he helped uncover critical vulnerabilities in AI systems. Currently at SentinelOne, he specializes in generative AI, AI alignment, and the benchmarking and measurement of AI system performance, focusing on Retrieval-Augmented Generation (RAG) and AI guardrails. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.sentinelone.com/
All the Hard Stuff with LLMs in Product Development // Phillip Carter // MLOps Podcast #170: https://www.youtube.com/watch?v=DZgXln3v85s&ab_channel=MLOps.community --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ron on LinkedIn: https://www.linkedin.com/in/heichmanron/
This is a panel taken from the recent AI quality Conference presented by the MLOps Community and Kolena
// Abstract The need for moving to production quickly is paramount in staying out of perpetual POC territory. AI is moving fast. Shipping features fast to stay ahead of the competition is commonplace. Quick iterations are viewed as strength in the startup ecosystem, especially when taking on a deeply entrenched competitor. Each week a new method to improve your AI system becomes popular or a SOTA foundation model is released. How do we balance the need for speed vs the responsibility of safety? Having the confidence to ship a cutting-edge model or AI architecture and knowing it will perform as tasked. What are the risks and safety metrics that others are using when they deploy their AI systems. How can you correctly identify when risks are too large? // Panelists - Remy Thellier: Head of Growth & Strategic Partnerships @ Vectice - Erica Greene: Director of Engineering, Machine Learning @ Yahoo - Shreya Rajpal: Creator @ Guardrails AI A big thank you to our Premium Sponsors Google Cloud & Databricks for their generous support!
Chinar Movsisyan is the co-founder and CEO of Feedback Intelligence (formerly Manot), an MLOps startup based in San Francisco. She has been in the AI field for more than 7 years from research labs to venture-backed startups.
Reliable LLM Products, Fueled by Feedback // MLOps Podcast #250 with Chinar Movsisyan, CEO of Feedback Intelligence. // Abstract We live in a world driven by large language models (LLMs) and generative AI, but ensuring they are ready for real-world deployment is crucial. Despite the availability of numerous evaluation tools, many LLM products still struggle to make it to production. We propose a new perspective on how LLM products should be measured, evaluated, and improved. A product is only as good as the user's experience and expectations, and we aim to enhance LLM products to meet these standards reliably. Our approach creates a new category that automates the need for separate evaluation, observability, monitoring, and experimentation tools. By starting with the user experience and working backward to the model, we provide a comprehensive view of how the product is actually used, rather than how it is intended to be used. This user-centric aka feedback-centric approach is the key to every successful product. // Bio Chinar Movsisyan is the founder and CEO of Feedback Intelligence, an MLOps company based in San Francisco that enables enterprises to make sure that LLM-based products are reliable and that the output is aligned with end-user expectations. With over eight years of experience in deep learning, spanning from research labs to venture-backed startups, Chinar has led AI projects in mission-critical applications such as healthcare, drones, and satellites. Her primary research interests include artificial intelligence, generative AI, machine learning, deep learning, and computer vision. At Feedback Intelligence, Chinar and her team address a crucial challenge in LLM development by automatically converting user feedback into actionable insights, enabling AI teams to analyze root causes, prioritize issues, and accelerate product optimization. This approach is particularly valuable in highly regulated industries, helping enterprises to reduce time-to-market and time-to-resolution while ensuring robust LLM products. Feedback Intelligence, which participated in the Berkeley SkyDeck accelerator program, is currently expanding its business across various verticals. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.manot.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Chinar on LinkedIn: https://www.linkedin.com/in/nik-suresh/ Timestamps: [00:00] Chinar's preferred coffee [00:20] Takeaways [02:25] Please like, share, leave a review, and subscribe to our MLOps channels! [03:23] Object Detection on Drones [06:10] Street Surveillance Detection Use Case [08:00] Optimizing Vision Models [09:50] Data Engineering for AI/ML Conference Ad [10:42] Plastic surgery project [12:33] Diffusion models getting popular [13:57] AI challenges in highly regulated industries [17:48] Product metrics evaluation insights [20:55] Chatbot effectiveness metrics [23:15] Interpreting user signals [24:45] Metadata tracking in LLM [27:41] Agentic workflow [28:53] Effective data analysis strategies [30:41] Identifying key metrics [33:59] AI metrics role shift [37:20] Tooling for non-engineers [42:12] Balancing engineering and evaluation [44:39] Bridging SME engineering gap [46:41] Expand expertise potential [47:40] What's with flamingos [48:04] Wrap up
This is a Panel taken from the recent AI Quality Conference presented by the MLOps COmmunity and Kolena
// Abstract Enterprise AI leaders continue to explore the best productivity solutions that solve business problems, mitigate risks, and increase efficiency. Building reliable and secure AI/ML systems requires following industry standards, an operating framework, and best practices that can accelerate and streamline the scalable architecture that can produce expected business outcomes. This session, featuring veteran practitioners, focuses on building scalable, reliable, and quality AI and ML systems for the enterprises. // Panelists - Hira Dangol: VP, AI/ML and Automation @ Bank of America - Rama Akkiraju: VP, Enterprise AI/ML @ NVIDIA - Nitin Aggarwal: Head of AI Services @ Google - Steven Eliuk: VP, AI and Governance @ IBM A big thank you to our Premium Sponsors Google Cloud & Databricks for their generous support!
Timestamps:
00:00 Panelists discuss vision and strategy in AI
05:18 Steven Eliuk, IBM expertise in data services
07:30 AI as means to improve business metrics
11:10 Key metrics in production systems: efficiency and revenue
13:50 Consistency in data standards aids data integration
17:47 Generative AI presents new data classification risks
22:47 Evaluating implications, monitoring, and validating use cases
26:41 Evaluating natural language answers for efficient production
29:10 Monitoring AI models for performance and ethics
31:14 AI metrics and user responsibility for future models
34:56 Access to data is improving, promising progress
Nik Suresh wrote an evisceration of the current AI hype boom called "I Will F**king Piledrive You If You Mention AI Again." AI Operations Without Fundamental Engineering Discipline // MLOps Podcast #250 with Nikhil Suresh, Director @ Hermit Tech. // Abstract Nik is on the podcast because of an anti-AI hype piece, so a reasonable thing to discuss is going to be what most companies are getting wrong when non-technical management wants to immediately roll out ML initiatives, but are unwilling to bring technical naysayers on board who will set them up for success. // Bio Nik is the author of ludic.mataroa.blog, who wrote "I Will [REDACTED] Piledriver You If You Mention AI Again", and mostly works in the data engineering and data science spaces. Nik's writing and company both focus on bringing more care to work, pushing back against the industry's worst excesses both technically and spiritually, and getting fundamentals right. Nik also has a reasonably strong background in psychology. His data science training was of the pre-LLM variety, circa. 2018, when there was a lot of hype but it wasn't this ridiculous. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://ludic.mataroa.blog/
Nik's blog: https://ludic.mataroa.blog/blog/i-will-fucking-piledrive-you-if-you-mention-ai-again/ Harnessing MLOps in Finance // Michelle Marie Conway // MLOps Podcast Coffee #174: https://youtu.be/nIEld_Q6L-0
Fundamentals of Data Engineering: Plan and Build Robust Data Systems AudiobookBy: Joe Reis, Matt Housley: https://audiobookstore.com/audiobooks/fundamentals-of-data-engineering.aspx
Bullshit Jobs A Theory Hardcover by David Graeber: https://www.amazon.co.jp/-/en/David-Graeber/dp/0241263883 Does a Frog have Scorpion Nature podcast: https://open.spotify.com/show/57i8sYVqxG4i3NvBniLfhv --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Nik on LinkedIn: https://www.linkedin.com/in/nik-suresh/
Timestamps: [00:00] Nik's preferred coffee [00:30] Takeaways [01:40] Please like, share, leave a review, and subscribe to our MLOps channels! [01:56] AI hype and humor [07:21] Defining project success [08:57] Effective data utilization [12:18] AI Hype vs Data Engineering [14:44] AI implementation challenges [17:44 - 18:35] Data Engineering for AI and ML Virtual Conference Ad [18:35] Managing AI Expectations [22:08] AI expectations vs reality [26:00] Balancing Engineering and AI [31:54] Highlighting engineer success [35:25] The real challenges [36:30] Embracing work challenges [37:21] Dealing with podcast disappointments [40:50] Creating content for visibility [43:02] Exploring niche interests [44:14] Relationship building [47:15] Strategic approach to success [48:36] Wrap up
Eric Landry is a seasoned AI and Machine Learning leader with extensive expertise in software engineering and practical applications in NLP, document classification, and conversational AI. With technical proficiency in Java, Python, and key ML tools, he leads the Expedia Machine Learning Engineering Guild and has spoken at major conferences like Applied Intelligence 2023 and KDD 2020. AI in Healthcare // MLOps Podcast #249 with Eric Landry, CTO/CAIO @ Zeteo Health. // Abstract Eric Landry discusses the integration of AI in healthcare, highlighting use cases like patient engagement through chatbots and managing medical data. He addresses benchmarking and limiting hallucinations in LLMs, emphasizing privacy concerns and data localization. Landry maintains a hands-on approach to developing AI solutions and navigating the complexities of healthcare innovation. Despite necessary constraints, he underscores the potential for AI to proactively engage patients and improve health outcomes. // Bio Eric Landry is a technology veteran with 25+ years of experience in the healthcare, travel, and computer industries, specializing in machine learning engineering and AI-based solutions. Holding a Masters in SWE (NLP thesis topic) from the University of Texas at Austin, 2005. He has showcased his expertise and leadership in the field with three US patents, published articles on machine learning engineering, and speaking engagements at the 2023 Applied Intelligence Live, 2020 KDD conference, Data Science Salon 2024, and former leader of Expedia’s MLE guild. Formerly, Eric was the director of AI Engineering and Conversation Platform at Babylon Health and Expedia. Currently CTO/CAIO at Zeteo Health. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.zeteo.health/ Building Threat Detection Systems: An MLE's Perspective // Jeremy Jordan // MLOps Podcast #134: https://youtu.be/13nOmMJuiAo --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Eric on LinkedIn: https://www.linkedin.com/in/jeric-landry/ Timestamps: [00:00] Eric's preferred coffee [00:16] Takeaways [01:16] Please like, share, leave a review, and subscribe to our MLOps channels! [01:32] ML and AI in 2005 [04:43] Last job at Babylon Health [10:57] Data access solutions [14:35] Prioritize AI ML Team Success [16:39] Eric's current work [20:36] Engage in holistic help [22:13] High-stakes chatbots [27:30] Navigating Communication Across Diverse Communities [31:49] When Bots Go Wrong [34:15] Health care challenges ahead [36:05] Behavioral health tech challenges [39:45] Stress from Apps Notifications [41:11] Combining different guardrails tools [47:16] Navigating Privacy AI [50:12] Wrap up
Aniket Kumar Singh is a Vision Systems Engineer at Ultium Cells, skilled in Machine Learning and Deep Learning. I'm also engaged in AI research, focusing on Large Language Models (LLMs). Evaluating the Effectiveness of Large Language Models: Challenges and Insights // MLOps Podcast #248 with Aniket Kumar Singh, CTO @ MyEvaluationPal | ML Engineer @ Ultium Cells. // Abstract Dive into the world of Large Language Models (LLMs) like GPT-4. Why is it crucial to evaluate these models, how we measure their performance, and the common hurdles we face? Drawing from Aniket's research, he shares insights on the importance of prompt engineering and model selection. Aniket also discusses real-world applications in healthcare, economics, and education, and highlights future directions for improving LLMs. // Bio Aniket is a Vision Systems Engineer at Ultium Cells, skilled in Machine Learning and Deep Learning. I'm also engaged in AI research, focusing on Large Language Models (LLMs). // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.aniketsingh.me Aniket's AI Research for Good blog that I plan to utilize to share any new research that would focus on the good: www.airesearchforgood.org Aniket's papers: https://scholar.google.com/citations?user=XHxdWUMAAAAJ&hl=en --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Aniket on LinkedIn: https://www.linkedin.com/in/singh-k-aniket/ Timestamps: [00:00] Aniket's preferred coffee [00:14] Takeaways [01:29] Aniket's job and hobby [03:06] Evaluating LLMs: Systems-Level Perspective [05:55] Rule-based system [08:32] Evaluation Focus: Model Capabilities [13:04] LLM Confidence [13:56] Problems with LLM Ratings [17:17] Understanding AI Confidence Trends [18:28] Aniket's papers [20:40] Testing AI Awareness [24:36] Agent Architectures Overview [27:05] Leveraging LLMs for tasks [29:53] Closed systems in Decision-Making [31:28] Navigating model Agnosticism [33:47] Robust Pipeline vs Robust Prompt [34:40] Wrap up
Sophia Rowland is a Senior Product Manager focusing on ModelOps and MLOps at SAS. In her previous role as a data scientist, Sophia worked with dozens of organizations to solve a variety of problems using analytics.
David Weik has a passion for data and creating integrated customer-centric solutions. Thinking data and people first to create value-added solutions. Extending AI: From Industry to Innovation // MLOps Podcast #247 with Sophia Rowland, Senior Product Manager and David Weik, Senior Solutions Architect of SAS. Huge thank you to SAS for sponsoring this episode. SAS - http://www.sas.com/ // Abstract Organizations worldwide invest hundreds of billions into AI, but they do not see a return on their investments until they are able to leverage their analytical assets and models to make better decisions. At SAS, we focus on optimizing every step of the Data and AI lifecycle to get high-performing models into a form and location where they drive analytically driven decisions. Join experts from SAS as they share learnings and best practices from implementing MLOps and LLMOPs at organizations across industries, around the globe, and using various types of models and deployments, from IoT CV problems to composite flows that feature LLMs. // Bio Sophia Rowland Sophia Rowland is a Senior Product Manager focusing on ModelOps and MLOps at SAS. In her previous role as a data scientist, Sophia worked with dozens of organizations to solve a variety of problems using analytics. As an active speaker and writer, Sophia has spoken at events like All Things Open, SAS Explore, and SAS Innovate as well as written dozens of blogs and articles. As a staunch North Carolinian, Sophia holds degrees from both UNC-Chapel Hill and Duke including bachelor’s degrees in computer science and psychology and a Master of Science in Quantitative Management: Business Analytics from the Fuqua School of Business. Outside of work, Sophia enjoys reading an eclectic assortment of books, hiking throughout North Carolina, and trying to stay upright while ice skating. David Weik David joined SAS in 2020 as a solutions architect. He helps customers to define and implement data-driven solutions. Previously, David was a SAS administrator/developer at a German insurance company working with the integration capabilities of SAS, Robotic Process Automation, and more. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links http://www.sas.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sophia on LinkedIn: https://www.linkedin.com/in/sophia-rowland/ Connect with David on LinkedIn: https://www.linkedin.com/in/david-weik/ Timestamps: [00:00] Sophia & David's preferred coffee [00:19] Takeaways [02:11] Please like, share, leave a review, and subscribe to our MLOps channels! [02:55] Hands on MLOps and AI [05:14] Next-Gen MLOps Challenges [07:24] Data scientists adopting software [11:48] Taking a different approach [13:43] Zombie Model Management [16:36] Optimizing ML Revenue Allocation [18:39] Other use cases - Lockout - Tagout procedure [21:43] Vision Model Integration Challenges [26:16] Costly errors in predictive maintenance [27:25] Integration of Gen AI [34:32] Governance challenges in AI [38:00] Governance in Gen AI vs Governance with Traditional ML [41:53] Evaluation challenges in industries [46:49] Interface frustration with Chatbots [51:25] Implementing AI Agent's success [54:18] Usability challenges in interfaces [57:03] Themes in High-Performing AI Teams [1:00:51] Wrap up
Matar Haller is the VP of Data & AI at ActiveFence, where her teams own the end-to-end automated detection of harmful content at scale, regardless of the abuse area or media type. The work they do here is engaging, impactful, and tough, and Matar is grateful for the people she gets to do it with.
AI For Good - Detecting Harmful Content at Scale // MLOps Podcast #246 with Matar Haller, VP of Data & AI at ActiveFence. // Abstract One of the biggest challenges facing online platforms today is detecting harmful content and malicious behavior. Platform abuse poses brand and legal risks, harms the user experience, and often represents a blurred line between online and offline harm. So how can online platforms tackle abuse in a world where bad actors are continuously changing their tactics and developing new ways to avoid detection? // Bio Matar Haller leads the Data & AI Group at ActiveFence, where her teams are responsible for the data, algorithms, and infrastructure that fuel ActiveFence’s ability to ingest, detect, and analyze harmful activity and malicious content at scale in an ever-changing, complex online landscape. Matar holds a Ph.D. in Neuroscience from the University of California at Berkeley, where she recorded and analyzed signals from electrodes surgically implanted in human brains. Matar is passionate about expanding leadership opportunities for women in STEM fields and has three children who surprise and inspire her every day. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links activefence.comhttps://www.youtube.com/@ActiveFence --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Matar on LinkedIn: https://www.linkedin.com/company/11682234/admin/feed/posts/ Timestamps: [00:00] Matar's preferred coffee [00:13] Takeaways [01:39] The talk that stood out [06:15] Online hate speech challenges [08:13] Evaluate harmful media API [09:58] Content moderation: AI models [11:36] Optimizing speed and accuracy [13:36] Cultural reference AI training [15:55] Functional Tests [20:05] Continuous adaptation of AI [26:43] AI detection concerns [29:12] Fine-Tuned vs Off-the-Shelf [32:04] Monitoring Transformer Model Hallucinations [34:08] Auditing process ensures accuracy [38:38] Testing strategies for ML [40:05] Modeling hate speech deployment [42:19] Improving production code quality [43:52] Finding balance in Moderation [47:23] Model's expertise: Cultural Sensitivity [50:26] Wrap up
Catherine Nelson is a freelance data scientist and writer. She is currently working on the forthcoming O’Reilly book "Software Engineering for Data Scientists”. Why All Data Scientists Should Learn Software Engineering Principles // MLOps podcast #245 with Catherine Nelson, a freelance Data Scientist. A big thank you to LatticeFlow AI for sponsoring this episode! LatticeFlow AI - https://latticeflow.ai/ // Abstract Data scientists have a reputation for writing bad code. This quote from Reddit sums up how many people feel: “It's honestly unbelievable and frustrating how many Data Scientists suck at writing good code.” But as data science projects grow, and because the job now often includes deploying ML models, it's increasingly important for DSs to learn fundamental SWE principles such as keeping your code modular, making sure your code is readable by other people and so on. The exploratory nature of DS projects means that you can't be sure where you will end up at the start of a project, but there's still a lot you can do to standardize the code you write. // Bio Catherine Nelson is the author of "Software Engineering for Data Scientists", a guide for data scientists who want to level up their coding skills, published by O'Reilly in May 2024. She is currently consulting for GenAI startups and providing mentorship and career coaching to data scientists. Previously, she was a Principal Data Scientist at SAP Concur. She has extensive experience deploying NLP models to production and evaluating ML systems, and she is also co-author of the book "Building Machine Learning Pipelines", published by O'Reilly in 2020. In her previous career as a geophysicist, she studied ancient volcanoes and explored for oil in Greenland. Catherine has a PhD in geophysics from Durham University and a Masters of Earth Sciences from Oxford University. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Software Engineering for Data Scientists book by Catherine Nelson: https://learning.oreilly.com/library/view/software-engineering-for/9781098136192/ https://www.amazon.com/Software-Engineering-Data-Scientists-Notebooks/dp/1098136209 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Catherine on LinkedIn: https://www.linkedin.com/in/catherinenelson1/ Timestamps: [00:00] Catherine's preferred coffee [00:15] Takeaways [02:38] Meeting magic: Embracing serenity [06:23] The Software Engineering for Data Scientists book [10:41] Exploring ideas rapidly [12:52] Bridging Data Science gaps [16:17] Data poisoning concerns [18:26] Transitioning from a data scientist to a machine learning engineer [21:53] Rapid Prototyping vs Thorough Development [23:45] Data scientists take ownership [25:53] Data scientists' role balance [30:30] Understanding system design process [36:00] Data scientists and Kubernetes [41:33 - 43:03] LatticeFlow AI Ad [43:05] The Future of Data Science [45:09] Data scientists analyzing models [46:46] Tools gaps in prompt tracking [50:44] Learnings from writing the book
Meta GenAI Infra Blog Review // Special MLOps Podcast episode by Demetrios. // Abstract Demetrios explores Meta's innovative infrastructure for large-scale AI operations, highlighting three blog posts on training large language models, maintaining AI capacity, and building Meta's GenAI infrastructure. The discussion reveals Meta's handling of hundreds of trillions of AI model executions daily, focusing on scalability, cost efficiency, and robust networking. Key elements include the Ops planner work orchestrator, safety protocols, and checkpointing challenges in AI training. Meta's efforts in hardware design, software solutions, and networking optimize GPU performance, with innovations like a custom Linux file system and advanced networking file systems like Hammerspace. The podcast also discusses advancements in PyTorch, network technologies like Roce and Nvidia's Quantum 2 Infiniband fabric, and Meta's commitment to open-source AGI. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Building Meta’s GenAI Infrastructure blog: https://engineering.fb.com/2024/03/12/data-center-engineering/building-metas-genai-infrastructure/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Timestamps:
[00:00] Meta handles trillions of AI model executions
[07:01] Meta creating AGI, ethical and sustainable
[08:13] Concerns about energy use in training models
[12:22] Network, hardware, and job optimization for reliability
[17:21] Highlights of Arista and Nvidia hardware architecture
[20:11] Meta's clusters optimized for efficient fabric
[24:40] Varied steps, careful checkpointing in AI training
[28:46] Meta is maintaining huge GPU clusters for AI
[29:47] AI training is faster and more demanding
[35:27] Ops planner orchestrates a million operations and reduces maintenance
[37:15] Ops planner ensures safety and well-tested changes
Sean Wei, the CEO and co-founder of RealChar, shares his journey from working in the autonomous vehicle industry to creating an open-source voice assistant project called Realchar, which eventually evolved into Rivia, a voice AI assistant focused on managing personal phone calls. The Future of AI and Consumer Empowerment // MLOps podcast #244 with Shaun Wei, CEO & Co-Founder of RealChar. A big thank you to LatticeFlow for sponsoring this episode! LatticeFlow - https://latticeflow.ai/ // Abstract Explore the groundbreaking work RealChar is doing with its consumer application, Rivia. This discussion focuses on how Rivia leverages Generative AI and Traditional Machine Learning to handle mundane phone calls and customer service interactions, aiming to free up human time for more meaningful tasks. The product, currently in beta, embodies a forward-thinking approach to AI, where the technology offloads day-to-day burdens like scheduling appointments and making calls. // Bio Shaun Wei is a well-connected technology professional with a rich background in developing and analyzing artificial intelligence systems. In 2018, Shaun played a pivotal role in the advent and deployment of Google Duplex, a remarkable AI capable of handling natural conversations and performing tasks such as booking hair salon appointments and restaurant reservations via telephone. His involvement wasn't just limited to the developmental side; Shaun also uniquely positioned himself on the receiving end, gathering insights by interviewing users directly impacted by the technology. This dual perspective has enabled Shaun to grasp both the technical underpinnings and the human-centric applications of AI, making him a valuable asset in the tech industry. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://www.rivia.tech/
https://realchar.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Shaun on LinkedIn: https://www.linkedin.com/in/shaunwei/
Timestamps: [00:00] Shaun's preferred coffee [00:28] Takeaways [03:30] Please like, share, leave a review, and subscribe to our MLOps channels! [03:57] AI in Production: Challenges & Insights [06:13] AI Scheduling and Assistance [08:00] Technical Challenges in AI [12:36 - 14:06] LatticeFlow Ad [14:09] Handling Challenges in AI [15:52] Learning driving and technical aspects [19:04] Self-Driving Cars: Multimodal Integration [23:41] Processing data with Transformers [26:46] Real-time phone data gathering [30:49] Real-time observability in AI [35:09] Time to first token [37:26] Preferred vs. Dynamic Model Selection [40:12] Event-driven architecture basics [42:06] Navigating challenges together [44:02] Challenges with Inconsistent Responses [45:40] Importance of product reliability [47:47] Training Data and Model Performance [50:02] Exploring AI in Customer Service [51:34] Navigating challenges in AI [53:15] Excited Launch Strategy Advice [57:10] Wrap up
Join us at our first in-person conference today all about AI Quality: https://www.aiqualityconference.com/ ML and AI as Distinct Control Systems in Heavy Industrial Settings // MLOps podcast #243 with Richard Howes, CTO of Metaformed.
Richard Howes is a dedicated engineer who is passionate about control systems whether it be embedded systems, industrial automation, or AI/ML in a business application. Huge thank you to AWS for sponsoring this episode. AWS - https://aws.amazon.com/ // Abstract How can we balance the need for safety, reliability, and robustness with the extreme pace of technology advancement in heavy industry? The key to unlocking the full potential of data will be to have a mixture of experts both from an AI and human perspective to validate anything from a simple KPI to a Generative AI Assistant guiding operators throughout their day. The data generated by heavy industries like agriculture, oil & gas, forestry, real estate, civil infrastructure, and manufacturing is underutilized and struggles to keep up with the latest and greatest - and for good reason. They provide the shelter we live and work in, the food we eat, and the energy to propel society forward. Compared to the pace of AI innovation they move slowly, have extreme consequences for failure, and typically involve a significant workforce. During this discussion, we will outline the data ready to be utilized by ML, AI, and data products in general as well as some considerations for creating new data products for these heavy industries. To account for complexity and uniqueness throughout the organization it is critical to engage operational staff, ensure safety is considered from all angles, and build adaptable ETL needed to bring the data to a usable state. // Bio Richard Howes is a dedicated engineer who is passionate about control systems whether it be embedded systems, industrial automation, or AI/ML in a business application. All of these systems require a robust control philosophy that outlines the system, its environment, and how the controller should function within it. Richard has a bachelor's of Electrical Engineering from the University of Victoria where he specialized in industrial automation and embedded systems. Richard is primarily focused on the heavy industrial sectors like energy generation, oil & gas, pulp/paper, forestry, real estate, and manufacturing. He works on both physical process control and business process optimization using the control philosophy principles as a guiding star. Richard has been working with industrial systems for over 10 years designing, commissioning, operating, and maintaining automated systems. For the last 5 years, Richard has been investing time into the data and data science-related disciplines bringing the physical process as close as possible to the business taking advantage of disparate data sets throughout the organization. Now with the age of AI upon us, he is focusing on integrating this technology safely, reliably, and with distinct organizational goals and ROI. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AWS Trainium: https://aws.amazon.com/machine-learning/trainium/ AWS Inferentia: https://aws.amazon.com/machine-learning/inferentia/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Richard on LinkedIn: https://www.linkedin.com/in/richardhowes/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ Accelerating Multimodal AI // MLOps podcast #241 with Ethan Rosenthal, Member of Technical Staff of Runway. Huge thank you to AWS for sponsoring this episode. AWS - https://aws.amazon.com/ // Abstract We’re still trying to figure out systems and processes for training and serving “regular” machine learning models, and now we have multimodal AI to contend with! These new systems present unique challenges across the spectrum, from data management to efficient inference. I’ll talk about the similarities, differences, and challenges that I’ve seen by moving from tabular machine learning, to large language models, to generative video systems. I’ll also talk about the setups and tools that I have seen work best for supporting and accelerating both the research and productionization process. // Bio Ethan works at Runway building systems for media generation. Ethan's work generally straddles the boundary between research and engineering without falling too hard on either side. Prior to Runway, Ethan spent 4 years at Square. There, he led a small team of AI Engineers training large language models for Conversational AI. Before Square, Ethan freelance consulted and worked at a couple ecommerce startups. Ethan found his way into tech by way of a Physics PhD. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.ethanrosenthal.com Ethan's mangum opus: https://www.ethanrosenthal.com/2020/08/25/optimal-peanut-butter-and-banana-sandwiches/ Real-time Model Inference in a Video Streaming Environment // Brannon Dorsey // Coffee Sessions #98: https://youtu.be/TNO6rYwP3yg Feature Stores for Self-Service Machine Learning: https://www.ethanrosenthal.com/2021/02/03/feature-stores-self-service/ Gen-1: The Next Step Forward for Generative AI: https://research.runwayml.com/gen1 Machine Learning: The High Interest Credit Card of Technical Debt by D. Sculley et al.: https://research.google/pubs/machine-learning-the-high-interest-credit-card-of-technical-debt/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ethan on LinkedIn: https://bsky.app/profile/ethanrosenthal.com
Timestamps: [00:00] Ethan's preferred coffee [00:11] Takeaways [02:07] Falling into LLMs [03:16] Advanced AI Tech Capabilities [04:40] AI-powered video editing tool [06:56] Transition to AI: Diffusion Models [09:09] Multimodal Feature Store breakdown [15:33] Multimodal Feature Stores Evolution [18:09] Benefits of Multimodal Feature Store [25:09] Centralized Training Data Repository [27:33] Large-scale distributed training [32:37 - 33:39] AWS Ad [33:45] Dealing with researchers on productionizing [43:52] Infrastructure for Researchers and Engineers [47:04] Generative DevOps movement [49:21] Structuring teams [52:06] Multimodal Feature Stores Efficiency [54:02] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ Navigating the AI Frontier: The Power of Synthetic Data and Agent Evaluations in LLM Development // MLOps podcast #241 with Boris Selitser, Co-Founder and CTO/CPO of Okareo. A big thank you to LatticeFlow for sponsoring this episode! LatticeFlow - https://latticeflow.ai/ // Abstract Explore the evolving landscape of building LLM applications, focusing on the critical roles of synthetic data and agent evaluations. Discover how synthetic data enhances model behavior description, prototyping, testing, and fine-tuning, driving robustness in LLM applications. Learn about the latest methods for evaluating complex agent-based systems, including RAG-based evaluations, dialog-level assessments, simulated user interactions, and adversarial models. This talk delves into the specific challenges developers face and the tradeoffs involved in each evaluation approach, providing practical insights for effective AI development. // Bio Boris is the Co-Founder and CTO/CPO at Okareo. Okareo is a full-cycle platform for developers to evaluate and customize AI/LLM applications. Before Okareo, Boris was Director of Product at Meta/Facebook, leading teams building internal platforms and ML products. Examples include a copyright classification system across the Facebook apps and an engagement platform for over 200K developers, 500K+ creators, and 12M+ Oculus users. Boris has a bachelor’s in Computer Science from UC Berkeley. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://docs.okareo.com/blog/data_loop https://docs.okareo.com/blog/agent_eval The Real E2E RAG Stack // Sam Bean // MLOps Podcast #217 - https://youtu.be/8uZst7pgOw0
RecSys at Spotify // Sanket Gupta // MLOps Podcast #232 - https://youtu.be/byH-ARJA4gk --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Boris on LinkedIn: https://www.linkedin.com/in/selitser/
Timestamps: [00:00] Boris' preferred coffee [00:37] Takeaways [02:32] Please like, share, leave a review, and subscribe to our MLOps channels! [02:48] Software Engineering and Data Science [06:01] AI Transformative Potential Explained [10:31] Prompt Injection Protection Strategies [17:03] Agent's metrics for Jira [24:11] Data and Metrics Evolution [27:54] Evaluation Focus Enhances Systems [31:22 - 32:52] LatticeFlow AD [32:55] Custom Evaluation and Synthetic Data [36:23] Synthetic data for expansion, evaluation, and map [41:06] Diverse agents' personalities for readiness [44:25] Agent functions [46:17] Optimizing Routing Agents [50:04] Adapting to tool output for decision-making [52:56] Agent framework evolution [55:41] Agent framework for delivering value [57:03] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ MLOps Coffee Sessions Special episode with LatticeFlow, How to Build Production-Ready AI Models for Manufacturing, fueled by our Premium Brand Partner, LatticeFlow. Deploying AI models in manufacturing involves navigating several technical challenges such as costly data acquisition, class imbalances, data shifts, leakage, and model degradation over time. How can you uncover the causes of model failures and prevent them effectively? This discussion covers practical solutions and advanced techniques to build resilient, safe, and high-performing AI systems in the manufacturing industry. // Bio Pavol Bielik Pavol earned his PhD at ETH Zurich, specializing in machine learning, symbolic AI, synthesis, and programming languages. His groundbreaking research earned him the prestigious Facebook Fellowship in 2017, representing the sole European recipient, along with the Romberg Grant in 2016. Following his doctorate, Pavol's passion for ensuring the safety and reliability of deep learning models led to the founding of LatticeFlow. Building on a more than a decade of research, Pavol and a dynamic team of researchers at LatticeFlow developed a platform that equips companies with the tools to deliver robust and high-performance AI models, utilizing automatic diagnosis and improvement of data and models. Aniket Singh Vision Systems Engineer AI Researcher Mohan Mahadevan Mohan Mahadevan is a seasoned technology leader with 25 years of experience in building computer vision (CV) and machine learning (ML) based products. Mohan has led teams to successfully deliver real world solutions spanning hardware, software, and AI based solutions in over 20 product families across a diverse range of domains, including Semiconductors, Robotics, Fintech, and Insuretech. Mohan Mahadevan has led global teams in the development of cutting-edge technologies across a range of disciplines including computer vision, machine learning, optical and hardware architectures, system design, computational optimization and more. Jürgen Weichenberger 20+ years of advanced analytics, data science, database design, architecture, and implementation on various platforms to solve Complex Industry Problems. Industrial Analytics is the fusion of manufacturing, production, reliability, integrity, quality, sales- and market-analytics and covering 10 Industries. By combining skills and experience, we are creating the next-generation AI & ML Solutions for our clients. Leveraging a unique formula which allows us to model some of the most challenging manufacturing problems while building, scaling, and enabling the end-user to leverage the next generation data products. The Strategy & Innoation Team at Schneider is specialising on Industrial-Grade Challenges where we are applying ML & AI methods to achieve state of the art results. Personally, I am driving my team and my own education to extend the limits of AI & ML beyond the current possible. I hold more than 15 patents and I am working on new innovations. I am working with our partner eco-system to enrich our accelerators with modern ML/AI techniques and integrating robotic equipment allows me to create next generation solutions. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://latticeflow.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Timestamps: [00:00] Demetrios' Intro [00:48] Announcements [01:57] Join us at our first in-person conference on June 25 all about AI Quality! [03:39] Speakers' intros [06:00] AI ML uncommon use cases [10:14] Challenges in Implementing AI and ML in Heavy Industries [11:41] Optimizing AI use cases [18:07] Moving from PoC to Production [20:53] Hybrid AI Integration for Safety [28:28] Training AI for Defect Variability [33:18] Challenges in AI Integration [35:39] Metrics for Evaluating Success [37:27] Challenges in AI Integration [44:39] Usage of LLMs [50:34] Fine-tuning AI Models [53:20] Trust Dynamics: TML vs LLM [55:23] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Miguel Fierro is a Principal Data Science Manager at Microsoft and holds a PhD in robotics.
From Robotics to Recommender Systems // MLOps Podcast #240 with Miguel Fierro, Principal Data Science Manager at Microsoft. Huge thank you to Zilliz for sponsoring this episode. Zilliz - https://zilliz.com/. // Abstract Miguel explains the limitations and considerations of applying ML in robotics, contrasting its use against traditional control methods that offer exactness, which ML approaches generally approximate. He discusses the integration of computer vision and machine learning in sports for player movement tracking and performance analysis, highlighting collaborations with European football clubs and the role of artificial intelligence in strategic game analysis, akin to a coach's perspective. // Bio Miguel Fierro is a Principal Data Science Manager at Microsoft Spain, where he helps customers solve business problems using artificial intelligence. Previously, he was CEO and founder of Samsamia Technologies, a company that created a visual search engine for fashion items allowing users to find products using images instead of words, and founder of the Robotics Society of Universidad Carlos III, which developed different projects related to UAVs, mobile robots, humanoid robots, and 3D printers. Miguel has also worked as a robotics scientist at Universidad Carlos III of Madrid (UC3M) and King’s College London (KCL) and has collaborated with other universities like Imperial College London and IE University in Madrid. Miguel is an Electrical Engineer by UC3M, PhD in robotics by UC3M in collaboration with KCL, and graduated from MIT Sloan School of Management. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://miguelgfierro.com GitHub: https://github.com/miguelgfierro/RecSys at Spotify // Sanket Gupta // MLOps Podcast #232 - https://youtu.be/byH-ARJA4gkRecommenders joins LF AI & Data as new Sandbox project: https://cloudblogs.microsoft.com/opensource/2023/10/10/recommenders-joins-lf-ai-data-as-new-sandbox-project/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Miguel on LinkedIn: https://www.linkedin.com/in/miguelgfierro/ Timestamps: [00:00] Miguel's preferred coffee [00:11] Takeaways [02:25] Robotics [10:44] Simpler solutions over ML [15:11] Robotics and Computer Vision [19:15] Basketball object detection [22:43 - 23:50] Zilliz Ad [23:51] Mr. Recommenders and Recommender systems' common patterns [31:35] Embeddings and Feature Stores [42:34] Experiment ROI for leadership [47:17] Hi ROI investments [51:13] LLMs in Recommender Systems [54:51] Wrap up
Uber's Michelangelo: Strategic AI Overhaul and Impact // MLOps podcast #239 with Demetrios Brinkmann. Huge thank you to Weights & Biases for sponsoring this episode. WandB Free Courses - http://wandb.me/courses_mlops // Abstract Uber's Michelangelo platform has evolved significantly through three major phases, enhancing its capabilities from basic ML predictions to sophisticated uses in deep learning and generative AI. Initially, Michelangelo 1.0 faced several challenges such as a lack of deep learning support and inadequate project tiering. To address these issues, Michelangelo 2.0 and subsequently 3.0 introduced improvements like support for Pytorch, enhanced model training, and integration of new technologies like Nvidia’s Triton and Kubernetes. The platform now includes advanced features such as a Genai gateway, robust compliance guardrails, and a system for monitoring model performance to streamline and secure AI operations at Uber. // Bio At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios constantly learns and engages in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether analyzing the best paths forward, overcoming obstacles, or building Lego houses with his daughter. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links From Predictive to Generative – How Michelangelo Accelerates Uber’s AI Journey blog post: https://www.uber.com/en-JP/blog/from-predictive-to-generative-ai/
Uber's Michelangelo: https://www.uber.com/en-JP/blog/michelangelo-machine-learning-platform/ The Future of Feature Stores and Platforms // Mike Del Balso & Josh Wills // MLOps Podcast # 186: https://youtu.be/p5F7v-w4EN0
Machine Learning Education at Uber // Melissa Barr & Michael Mui // MLOps Podcast #156: https://youtu.be/N6EbBUFVfO8 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Timestamps: [00:00] Uber's Michelangelo platform evolution analyzed in podcast [03:51 - 4:50] Weights & Biases Ad [05:57] Uber creates Michelangelo to streamline machine learning [07:44] Michelangelo platform's tech and flexible system [11:49] Uber Michelangelo platform adapted for deep learning [16:48] Uber invests in ML training for employees [19:08] Explanation of blog content, ML quality metrics [22:38] Michelangelo 2.0 prioritizes serving latency and Kubernetes [26:30] GenAI gateway manages model routing and costs [31:35] ML platform evolution, legacy systems, and maintenance [33:22] Team debates maintaining outdated tools or moving on [34:41] Please like, share, leave feedback, and subscribe to our MLOps channels! [34:57] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Matthew McClean is a Machine Learning Technology Leader with the leading Amazon Web Services (AWS) cloud platform. He leads the customer engineering teams at Annapurna ML helping customers adopt AWS Trainium and Inferentia for their Gen AI workloads. Kamran Khan, Sr Technical Business Development Manager for AWS Inferentina/Trianium at AWS. He has over a decade of experience helping customers deploy and optimize deep learning training and inference workloads using AWS Inferentia and AWS Trainium. AWS Tranium and Inferentia // MLOps podcast #238 with Kamran Khan, BD, Annapurna ML and Matthew McClean, Annapurna Labs Lead Solution Architecture at AWS. Huge thank you to AWS for sponsoring this episode. AWS - https://aws.amazon.com/ // Abstract Unlock unparalleled performance and cost savings with AWS Trainium and Inferentia! These powerful AI accelerators offer MLOps community members enhanced availability, compute elasticity, and energy efficiency. Seamlessly integrate with PyTorch, JAX, and Hugging Face, and enjoy robust support from industry leaders like W&B, Anyscale, and Outerbounds. Perfectly compatible with AWS services like Amazon SageMaker, getting started has never been easier. Elevate your AI game with AWS Trainium and Inferentia! // Bio Kamran Khan Helping developers and users achieve their AI performance and cost goals for almost 2 decades. Matthew McClean Leads the Annapurna Labs Solution Architecture and Prototyping teams helping customers train and deploy their Generative AI models with AWS Trainium and AWS Inferentia // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AWS Trainium: https://aws.amazon.com/machine-learning/trainium/ AWS Inferentia: https://aws.amazon.com/machine-learning/inferentia/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Kamran on LinkedIn: https://www.linkedin.com/in/kamranjk/ Connect with Matt on LinkedIn: https://www.linkedin.com/in/matthewmcclean/ Timestamps: [00:00] Matt's & Kamran's preferred coffee [00:53] Takeaways [01:57] Please like, share, leave a review, and subscribe to our MLOps channels! [02:22] AWS Trainium and Inferentia rundown [06:04] Inferentia vs GPUs: Comparison [11:20] Using Neuron for ML [15:54] Should Trainium and Inferentia go together? [18:15] ML Workflow Integration Overview [23:10] The Ec2 instance [24:55] Bedrock vs SageMaker [31:16] Shifting mindset toward open source in enterprise [35:50] Fine-tuning open-source models, reducing costs significantly [39:43] Model deployment cost can be reduced innovatively [43:49] Benefits of using Inferentia and Trainium [45:03] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/.
Benjamin Wilms is a developer and software architect at heart, with 20 years of experience. He fell in love with chaos engineering. Benjamin now spreads his enthusiasm and new knowledge as a speaker and author – especially in the field of chaos and resilience engineering. Retrieval Augmented Generation // MLOps podcast #237 with Benjamin Wilms, CEO & Co-Founder of Steadybit. Huge thank you to Amazon Web Services for sponsoring this episode. AWS - https://aws.amazon.com/ // Abstract How to build reliable systems under unpredictable conditions with Chaos Engineering. // Bio Benjamin has over 20 years of experience as a developer and software architect. He fell in love with chaos engineering 7 years ago and shares his knowledge as a speaker and author. In October 2019, he founded the startup Steadybit with two friends, focusing on developers and teams embracing chaos engineering. He relaxes by mountain biking when he's not knee-deep in complex and distributed code. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://steadybit.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Benjamin on LinkedIn: https://www.linkedin.com/in/benjamin-wilms/ Timestamps: [00:00] Benjamin's preferred coffee [00:28] Takeaways [02:10] Please like, share, leave a review, and subscribe to our MLOps channels! [02:53] Chaos Engineering tldr [06:13] Complex Systems for smaller Startups [07:21] Chaos Engineering benefits [10:39] Data Chaos Engineering trend [15:29] Chaos Engineering vs ML Resilience [17:57 - 17:58] AWS Trainium and AWS Infecentia Ad [19:00] Chaos engineering tests system vulnerabilities and solutions
[23:24] Data distribution issues across different time zones
[27:07] Expertise is essential in fixing systems
[31:01] Chaos engineering integrated into machine learning systems
[32:25] Pre-CI/CD steps and automating experiments for deployments
[36:53] Chaos engineering emphasizes tool over value
[38:58] Strong integration into observability tools for repeatable experiments
[45:30] Invaluable insights on chaos engineering
[46:42] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Tom Smoker is the cofounder of an early stage tech company empowering developers to create knowledge graphs within their RAG pipelines. Tom is a technical founder, and owns the research and development of knowledge graphs tooling for the company. Managing Small Knowledge Graphs for Multi-agent Systems // MLOps podcast #236 with Tom Smoker, Technical Founder of whyhow.ai. A big thank you to @latticeflow for sponsoring this episode! LatticeFlow - https://latticeflow.ai/ // Abstract RAG is one of the more popular use cases for generative models, but there can be issues with repeatability and accuracy. This is especially applicable when it comes to using many agents within a pipeline, as the uncertainty propagates. For some multi-agent use cases, knowledge graphs can be used to structurally ground the agents and selectively improve the system to make it reliable end to end. // Bio Technical Founder of WhyHow.ai. Did Masters and PhD in CS, specializing in knowledge graphs, embeddings, and NLP. Worked as a data scientist to senior machine learning engineer at large resource companies and startups. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models: https://arxiv.org/abs/2401.01313Understanding the type of Knowledge Graph you need — Fixed vs Dynamic Schema/Data: https://medium.com/enterprise-rag/understanding-the-type-of-knowledge-graph-you-need-fixed-vs-dynamic-schema-data-13f319b27d9e --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Tom on LinkedIn: https://www.linkedin.com/in/thomassmoker/ Timestamps: [00:00] Tom's preferred coffee [00:33] Takeaways [03:04] Please like, share, leave a review, and subscribe to our MLOps channels! [03:23] Academic Curiosity and Knowledge Graphs [05:07] Logician [05:53] Knowledge graphs incorporated into RAGs [07:53] Graphs & Vectors Integration [10:49] "Exactly wrong" [12:14] Data Integration for Robust Knowledge Graph [14:53] Structured and Dynamic Data [21:44] Scoped Knowledge Retrieval Strategies [28:01 - 29:32] LatticeFlow Ad [29:33] RAG Limitations and Solutions [36:10] Working on multi agents, questioning agent definition
[40:01] Concerns about performance of agent information transfer
[43:45] Anticipating agent-based systems with modular processes
[52:04] Balancing risk tolerance in company operations and control
[54:11] Using AI to generate high-quality, efficient content
[01:03:50] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
David Nunez, based in Santa Barbara, CA, US, is currently a Co-Founder and Partner at Abstract Group, bringing experience from previous roles at First Round Capital, Stripe, and Slab. Just when we Started to Solve Software Docs, AI Blew Everything Up // MLOps Podcast #235 with Dave Nunez, Partner of Abstract Group co-hosted by Jakub Czakon. Huge thank you to Zilliz for sponsoring this episode. Zilliz - https://zilliz.com/. // Abstract Over the previous decade, the recipe for making excellent software docs mostly converged on a set of core goals: Create high-quality, consistent content Use different content types depending on the task Make the docs easy to find For AI-focused software and products, the entire developer education playbook needs to be rewritten. // Bio Dave lives in Santa Barbara, CA with his wife and four kids. He started his tech career at various startups in Santa Barbara before moving to San Francisco to work at Salesforce. After Salesforce, he spent 2+ years at Uber and 5+ years at Stripe leading internal and external developer documentation efforts. In 2021, he co-authored Docs for Developers to help engineers become better writers. He's now a consultant, advisor, and angel investor for fast-growing startups. He typically invests in early-stage startups focusing on developer tools, productivity, and AI. He's a reading nerd, Lakers fan, and golf masochist. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.abstractgroup.co/ Book: docsfordevelopers.com About Dave: https://gamma.app/docs/Dave-Nunez-about-me-002doxb23qbblme?mode=doc https://review.firstround.com/investing-in-internal-documentation-a-brick-by-brick-guide-for-startups https://increment.com/documentation/why-investing-in-internal-docs-is-worth-it/
Writing to Learn paper by Peter Elbow: https://peterelbow.com/pdfs/Writing_for_Learning-Not_just_Demonstrating.PDF --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Dave on LinkedIn: https://www.linkedin.com/in/djnunez/ Connect with Kuba on LinkedIn: https://www.linkedin.com/in/jakub-czakon/?locale=en_US Timestamps: [00:00] Dave's preferred coffee [00:13] Introducing this episode's co-host, Kuba [00:36] Takeaways [02:55] Please like, share, leave a review, and subscribe to our MLOps channels! [03:23] Good docs, bad docs, and how to feel them [06:51] Inviting Dev docs and checks [10:36] Stripe's writing culture [12:42] Engineering team writing culture [14:15] Bottom-up tech writer change [18:31] Strip docs cult following [24:40] TriDocs Smart API Injection [26:42] User research for documentation [29:51] Design cues [32:15] Empathy-driven docs creation [34:28 - 35:35] Zilliz Ad [35:36] Foundational elements in documentation [38:23] Minimal infrastructure of information in "Read Me" [40:18] Measuring documentation with OKRs [43:58] Improve pages with Analytics [47:33] Google branded doc searches [48:35] Time to First Action [52:52] Dave's day in and day out and what excites him [56:01] Exciting internal documentation [59:55] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Cody Peterson has a diverse work experience in the field of product management and engineering. Cody is currently working as a Technical Product Manager at Voltron Data, starting from May 2023. Previously, they worked as a Product Manager at dbt Labs from July 2022 to March 2023. MLOps podcast #234 with Cody Peterson, Senior Technical Product Manager at Voltron Data | Ibis project // Open Standards Make MLOps Easier and Silos Harder. Huge thank you to Weights & Biases for sponsoring this episode. WandB Free Courses -http://wandb.me/courses_mlops // Abstract MLOps is fundamentally a discipline of people working together on a system with data and machine learning models. These systems are already built on open standards we may not notice -- Linux, git, scikit-learn, etc. -- but are increasingly hitting walls with respect to the size and velocity of data. Pandas, for instance, is the tool of choice for many Python data scientists -- but its scalability is a known issue. Many tools make the assumption of data that fits in memory, but most organizations have data that will never fit in a laptop. What approaches can we take? One emerging approach with the Ibis project (created by the creator of pandas, Wes McKinney) is to leverage existing "big" data systems to do the heavy lifting on a lightweight Python data frame interface. Alongside other open source standards like Apache Arrow, this can allow data systems to communicate with each other and users of these systems to learn a single data frame API that works across any of them. Open standards like Apache Arrow, Ibis, and more in the MLOps tech stack enable freedom for composable data systems, where components can be swapped out allowing engineers to use the right tool for the job to be done. It also helps avoid vendor lock-in and keep costs low. // Bio Cody is a Senior Technical Product Manager at Voltron Data, a next-generation data systems builder that recently launched an accelerator-native GPU query engine for petabyte-scale ETL called Theseus. While Theseus is proprietary, Voltron Data takes an open periphery approach -- it is built on and interfaces through open standards like Apache Arrow, Substrait, and Ibis. Cody focuses on the Ibis project, a portable Python dataframe library that aims to be the standard Python interface for any data system, including Theseus and over 20 other backends. Prior to Voltron Data, Cody was a product manager at dbt Labs focusing on the open source dbt Core and launching Python models (note: models is a confusing term here). Later, he led the Cloud Runtime team and drastically improved the efficiency of engineering execution and product outcomes. Cody started his carrer as a Product Manager at Microsoft working on Azure ML. He spent about 2 years on the dedicated MLOps product team, and 2 more years on various teams across the ML lifecycel including data, training, and inferencing. He is now passionate about using open source standards to break down the silos and challenges facing real world engineering teams, where engineering increasingly involves data and machine learning. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Ibis Project: https://ibis-project.org Apache Arrow and the “10 Things I Hate About pandas”: https://wesmckinney.com/blog/apache-arrow-pandas-internals/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Cody on LinkedIn: https://linkedin.com/in/codydkdc
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Syed Asad is an Innovator, Generative AI & Machine Learning Engineer, and a Champion for Ethical AI
MLOps podcast #233 with Syed Asad, Lead AI/ML Engineer at KiwiTech // Retrieval Augmented Generation.
A big thank you to @ for sponsoring this episode! AWS -
// Abstract
Everything and anything around RAG.
// Bio
Currently Exploring New Horizons:
Syed is diving deep into the exciting world of Semantic Vector Searches and Vector Databases. These innovative technologies are reshaping how we interact with and interpret vast data landscapes, opening new avenues for discovery and innovation.
Specializing in Retrieval Augmented Generation (RAG):
Syed's current focus also includes mastering Retrieval Augmented Generation Techniques (RAGs). This cutting-edge approach combines the power of information retrieval with generative models, setting new benchmarks in AI's capability and application.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://sanketgupta.substack.com/
Our paper on this topic "Generalized User Representations for Transfer Learning": https://arxiv.org/abs/2403.00584
Sanket's blogs on Medium in the past: https://medium.com/@sanket107
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Syed on LinkedIn: https://www.linkedin.com/in/syed-asad-76815246/
Timestamps:
[00:00] Syed's preferred coffee
[00:31] Takeaways
[03:17] Please like, share, leave a review, and subscribe to our MLOps channels!
[03:37] A production issue
[07:37] CSV file handling risks
[09:42] Embedding models not suitable
[11:22] Inference layer experiments and use cases
[14:00] AWS service handling the issue
[17:35] Salad testing and insights
[22:12] OpenAI vs Customization
[24:30] Difference between Olama and VLLM
[27:16] Fine-tuning of small LLMs
[29:51] Evaluation framework
[32:04] MLOps for efficient ML
[37:12] Determining the pricing of tools
[39:35] Manage Dependency Risk
[40:27] Get in touch with Syed on LinkedIn
[41:46] ML Engineers are now all AI Engineers
[43:01] The hard framework
[43:53] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Sanket works as a Senior Machine Learning Engineer at Spotify working on building end-to-end audio recommender systems. Models built by his team are used across Spotify in many different products including Discover Weekly and Autoplay. MLOps podcast #232 with Sanket Gupta, Senior Machine Learning Engineer at Spotify // RecSys at Spotify. A big thank you to LatticeFlow for sponsoring this episode! LatticeFlow - https://latticeflow.ai/ // Abstract LLMs with foundational embeddings have changed the way we approach AI today. Instead of re-training models from scratch end-to-end, we instead rely on fine-tuning existing foundation models to perform transfer learning. Is there a similar approach we can take with recommender systems? In this episode, we can talk about: a) how Spotify builds and maintains large-scale recommender systems, b) how foundational user and item embeddings can enable transfer learning across multiple products, c) how we evaluate this system d) MLOps challenges with these systems // Bio Sanket works as a Senior Machine Learning Engineer on a team at Spotify building production-grade recommender systems. Models built by my team are being used in Autoplay, Daily Mix, Discover Weekly, etc. Currently, my passion is how to build systems to understand user taste - how do we balance long-term and short-term understanding of users to enable a great personalized experience. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://sanketgupta.substack.com/ Our paper on this topic "Generalized User Representations for Transfer Learning": https://arxiv.org/abs/2403.00584 Sanket's blogs on Medium in the past: https://medium.com/@sanket107 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sanket on LinkedIn: www.linkedin.com/in/sanketgupta107 Timestamps: [00:00] Sanket's preferred coffee [00:37] Takeaways [02:30] RecSys are RAGs [06:22] Evaluating RecSys parallel to RAGs [07:13] Music RecSys Optimization [09:46] Dealing with cold start problems [12:18] Quantity of models in the recommender systems [13:09] Radio models [16:24] Evaluation system [20:25] Infrastructure support [21:25] Transfer learning [23:53] Vector database features [25:31] Listening History Balance [26:35 - 28:06] LatticeFlow Ad [28:07] The beauty of embeddings [30:13] Shift to real-time recommendation [34:05] Vector Database Architecture Options [35:30] Embeddings drive personalized [40:16] Feature Stores vs Vector Databases [42:33] Spotify product integration strategy [45:38] Staying up to date with new features [47:53] Speed vs Relevance metrics [49:40] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Ryan Carson. CEO Founder for 20 years Built and sold 3 startups Helping build a global community of AI devs with Intel. MLOps podcast #231 with Ryan Carson, Senior AI Dev Community Lead at Intel Huge thank you to Zilliz for sponsoring this episode. Zilliz - https://zilliz.com/ // Abstract Ryan shares his professional journey, tracing his transition from building Treehouse to joining Intel. The conversation evolves into a deep dive into Carson's aspiration to democratize access to AI development. Furthermore, he expounds on the exciting prospects of new technology like Gaudi three, a new ASIC for AI workloads. Ryan emphasizes the need for driving competition in compute to lower prices and increase access, underlining the importance of associating individual work with company-based OKRs or KPIs. There is also a reflection on the essentiality of forging quality relationships in professional settings and aligning work with top-level OKRs. Discussion on the potential benefits of AI in constructing and maintaining professional interactions is explored. Touching upon practical applications of AI, they also delve into smaller projects, the possibility of one-person companies, and the role of AI for daily interactions. The episode concludes with an expression of optimism about technological advances shaping the future and an appreciation for the enlightening conversation. // Bio Ryan has been a founder, entrepreneur, and CEO for 20 years, successfully building, scaling, and selling three companies. He's passionate about empowering people to become developers and then connecting them together in a global community. After earning a degree in Computer Science in Colorado, Ryan moved to the UK and worked as a web developer. He then organized global tech conferences, hosting thousands of attendees and influential speakers such as Mark Zuckerberg, the founders of Android, Instagram, and Twitter, among others. His company also produced Twitter’s and Stack Overflow’s developer conferences. Following that, Ryan started an online Computer Science school. Under his leadership, the team grew to over 100 employees, educating more than 1,000,000 students. During this period, he secured $23 million in venture capital and earned recognition as Entrepreneur of the Year. Over the last two years Ryan dove deep into AI and LLMs. He built an educational proof-of-concept called maple.coach, which focuses on teaching Sales. The platform is built using technologies like Next.js, TypeScript, gpt-4, and Vercel. Outside of work, Ryan shares his life with his wife of 20 years and their two teenagers in Connecticut. They enjoy spending their free time sailing and taking walks with their Sheltie, Brinkley. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: ryancarson.com --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ryan on LinkedIn: https://www.linkedin.com/in/ryancarson/
Salman Avestimehr is a Dean's Professor, the inaugural director of the USC-Amazon Center for Secure and Trusted Machine Learning (Trusted AI), and director of the Information Theory and Machine Learning (vITAL) research lab. He is also the CEO and co-founder of FedML. MLOps podcast #230 with Salman Avestimehr, CEO & Founder of FedML, FedML Nexus AI: Your Generative AI Platform at Scale.
A big thank you to FEDML for sponsoring this episode! // Abstract FedML is your generative AI platform at scale to enable developers and enterprises to build and commercialize their own generative AI applications easily, scalably, and economically. Its flagship product, FedML Nexus AI, provides unique features in enterprise AI platforms, model deployment, model serving, AI agent APIs, launching training/Inference jobs on serverless/decentralized GPU cloud, experimental tracking for distributed training, federated learning, security, and privacy. // Bio Salman is a professor, the inaugural director of the USC-Amazon Center for Secure and Trusted Machine Learning (Trusted AI), and the director of the Information Theory and Machine Learning (vITAL) research lab at the Electrical and Computer Engineering Department and Computer Science Department of the University of Southern California. Salman is also the co-founder and CEO of FedML. He received his Ph.D. in Electrical Engineering and Computer Sciences from UC Berkeley in 2008. Salman does research in the areas of information theory, decentralized and federated machine learning, secure and privacy-preserving learning, and computing. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://www.avestimehr.com/ https://fedml.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Salman on LinkedIn: https://www.linkedin.com/company/fedml/ Timestamps: [00:00] AI Quality: First in-person conference on June 25 [01:28] Salman's preferred coffee [01:49] Takeaways [03:33] Please like, share, leave a review, and subscribe to our MLOps channels! [03:53] Challenges that inspired Salman's work [06:20] Controlled ownership [08:11] Dealing with data leakage and privacy problems [10:45] In-house ML Model Deployment [13:36] FEDML: Comprehensive Model Deployment [17:27] Integrating FEDML with Kubernetes [19:46] AI Evaluation Trends [24:37] Enhancing NLP with ML [25:48] FEDML: Canary, A/B, Confidence [29:36] FEDML customers [33:21] On-premise platform for secure data management
[37:16] Future prediction: data's crucial for better applications
[38:18] Maturity in evaluating and improving steps
[41:38] Focus on ownership
[45:12] Benefits of smaller models for specific use cases
[48:57] Verify sensitive tasks, trust quick, important mobile content creation
[51:50] Wrap up
Mohamed Elgendy is the Co-Founder & CEO at Kolena. Additionally, Mohamed Elgendy has had 1 past job as the Director Of Product and Engineering at Synapse Technology Corporation. Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ MLOps podcast #228 with Mohamed Elgendy, Co-founder & CEO of Kolena Inc., What is AI Quality? // Abstract Delve into the multifaceted concept of AI Quality. Demetrios and Mo explore the idea that AI quality is dependent on the specific domain, equitable to the difference in desired qualities between a $1 pen and a $100 pen. Mo underscores the performance of a product being in sync with its intended functionality and the absence of unknown risks as the pillars of AI Quality. They emphasize the need for comprehensive quality checks and adaptability of standards to differing product traits. Issues affecting edge deployments like latency are also highlighted. A deep dive into the formation of gold standards for AI, the nuanced necessities for various use cases, and the paramount need for collaboration among AI builders, regulators, and infrastructure firms form the core of the discussion. Elgendy brings to light their ambitious AI Quality Conference, aiming to set tangible, effective, but innovation-friendly Quality standards for AI. The dialogue also accentuates the urgent need for diversification and representation in the tech industry, the variability of standards and regulations, and the pivotal role of testing in AI and machine learning. The episode concludes with an articulate portrayal of how enhanced testing can streamline the entire process of machine learning. // Bio Mohamed is the Co-founder & CEO of Kolena and the author of the book “Deep Learning for Vision Systems”. Previously, he built and managed AI/ML organizations at Amazon, Twilio, Rakuten, and Synapse. Mohamed regularly speaks at AI conferences like Amazon's DevCon, O'Reilly's AI conference, and Google's I/O. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.kolena.io Deep Learning for Vision Systems book: https://www.amazon.com/Learning-Vision-Systems-Mohamed-Elgendy/dp/1617296198/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Mo on LinkedIn: https://www.linkedin.com/in/moelgendy/ Timestamps: [00:00] Mo's preferred coffee [00:07] Takeaways [02:52] See you all in San Francisco on June 25! [03:04] Please like, share, leave a review, and subscribe to our MLOps channels! [03:22] AI Quality in Mo's eyes [08:36] Quality Standards for Software [14:11] Common Chatbot Functionality [19:20] The Birth of Innovation [24:27] Transforming Insights into Standards [30:27] Testing: One step to quality [34:58] Two different data points to be harmonized [37:29] Model cards [39:12] Test Coverage Democratizes Collaboration [42:55] Representation matters [44:50] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com
Simon Karasik is a proactive and curious ML Engineer with 5 years of experience. Developed & deployed ML models at WEB and Big scale for Ads and Tax. Huge thank you to Nebius AI for sponsoring this episode. Nebius AI - https://nebius.ai/ MLOps podcast #228 with Simon Karasik, Machine Learning Engineer at Nebius AI, Handling Multi-Terabyte LLM Checkpoints. // Abstract The talk provides a gentle introduction to the topic of LLM checkpointing: why is it hard, how big are the checkpoints. It covers various tips and tricks for saving and loading multi-terabyte checkpoints, as well as the selection of cloud storage options for checkpointing. // Bio Full-stack Machine Learning Engineer, currently working on infrastructure for LLM training, with previous experience in ML for Ads, Speech, and Tax. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Simon on LinkedIn: https://www.linkedin.com/in/simon-karasik/ Timestamps: [00:00] Simon preferred beverage [01:23] Takeaways [04:22] Simon's tech background [08:42] Zombie models garbage collection [10:52] The road to LLMs [15:09] Trained models Simon worked on [16:26] LLM Checkpoints [20:36] Confidence in AI Training [22:07] Different Checkpoints [25:06] Checkpoint parts [29:05] Slurm vs Kubernetes [30:43] Storage choices lessons [36:02] Paramount components for setup [37:13] Argo workflows [39:49] Kubernetes node troubleshooting [42:35] Cloud virtual machines have pre-installed mentoring [45:41] Fine-tuning [48:16] Storage, networking, and complexity in network design [50:56] Start simple before advanced; consider model needs. [53:58] Join us at our first in-person conference on June 25 all about AI Quality
Sol Rashidi is an esteemed executive, leader, and influencer within the AI, Data, and Technology space. Having helped IBM launch Watson in 2011 as one of the earliest world applications of Artificial Intelligence, Sol has pioneered some of the early advancements of space. Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ Huge thank you to @WeightsBiases for sponsoring this episode. WandB Free Courses - http://wandb.me/courses_mlops MLOps podcast #227 with Sol Rashidi, CEO & Co-Founder of ExecutiveAI, Leading Enterprise Data Teams. // Abstract In the dynamic landscape of MLOps and data leadership, Sol shares invaluable insights on building successful teams and driving impactful projects. In this podcast episode, Sol delves into the importance of prioritizing relationships, introduces a pragmatic "Wrong Use Cases Formula" to streamline project prioritization, and emphasizes the critical role of effective communication in data leadership. Her wealth of experience and practical advice provide a roadmap for navigating the complexities of MLOps and leading data-driven initiatives to success. // Bio With eight (8) patents granted, 21 filed, and received awards that include: "Top 100 AI People" 2023 "The Top 75 Innovators of 2023" "Top 65 Most Influential Women in 2023" "Forbes AI Maverick of the 21st Century" 2022 “Top 10 Global Women in AI & Data”, 2023 "Top AI 100 Award", 2023 “50 Most Powerful Women in Tech”, 2022 “Global 100 Power List” - 2021, 2022, 2023 “Top 20 CDOs Globally” - 2022 "Chief Analytics Officer of the Year" - 2022 "Isomer Innovators of the Year" - 2021, 2022, 2023 "Top 100 Innovators in Data & Analytics” - 2020, 2021, 2022, 2023 "Top 100 Women in Business" - 2022 Sol is an energetic business executive and a goal-oriented technologist, skilled at coupling her technical acumen with story-telling abilities to articulate business value with both startups and Fortune 100's who are leaning into data, AI, and technology as a competitive advantage while wanting to preserve the legacy in which they were founded upon. Sol has served as a C-Suite member across several Fortune 100 & Fortune 500 companies including: Chief Analytics Officer - Estee Lauder Chief Data & Analytics Officer - Merck Pharmaceuticals EVP, Chief Data Officer - Sony Music Chief Data & AI Officer - Royal Caribbean Cruise Lines Sr. Partner leading the Digital & Innovation Practice- Ernsty & Young Partner leading Watson Go-To-Market & Commercialization - IBM Sol now serves as the CEO of ExecutiveAI LLC. A company dedicated to democratizing Artificial Intelligence for Humanity and is considered an outstanding and influential business leader who is influencing the space traveling the world as a keynote speaker, and serving as the bridge between established Gen1.0 markets and those evolving into 4.0. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Sol's Book will be out on April 30, 2024 Your AI Survival Guide: Scraped Knees, Bruised Elbows, and Lessons Learned from Real-World AI Deployments: https://www.amazon.com/Your-Survival-Guide-Real-World-Deployments/dp/1394272634?ref_=ast_author_mpb --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sol on LinkedIn: https://www.linkedin.com/in/sol-rashidi-a672291/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Chad Sanderson is passionate about data quality, and fixing the muddy relationship between data producers and consumers. He is a former Head of Data at Convoy, a LinkedIn writer, and a published author. He lives in Seattle, Washington, and is the Chief Operator of the Data Quality Camp. Huge thank you to @amazonwebservices for sponsoring this episode. AWS - https://aws.amazon.com/ MLOps podcast #226 with Chad Sanderson, CEO & Co-Founder of Gable, The Rise of Modern Data Management. // Abstract In this session, Chad Sanderson, CEO of Gable.ai and author of the upcoming O’Reilly book: "Data Contracts," tackles the necessity of modern data management in an age of hyper iteration, experimentation, and AI. He will explore why traditional data management practices fail and how the cloud has fundamentally changed data development. The talk will cover a modern application of data management best practices, including data change detection, data contracts, observability, and CI/CD tests, and outline the roles of data producers and consumers. Attendees will leave with a clear understanding of modern data management's components and how to leverage them for better data handling and decision-making. // Bio Chad Sanderson, CEO of Gable.ai, is a prominent figure in the data tech industry, having held key data positions at leading companies such as Convoy, Microsoft, Sephora, Subway, and Oracle. He is also the author of the upcoming O'Reilly book, "Data Contracts” and writes about the future of data infrastructure, modeling, and contracts in his newsletter “Data Products.” // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AWS Trainium and Inferentia: https://aws.amazon.com/machine-learning/trainium/ https://aws.amazon.com/machine-learning/inferentia/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Chad on LinkedIn: https://www.linkedin.com/in/chad-sanderson/
Patrick Beukema has a Ph.D. in neuroscience and has worked on AI models for brain decoding, which analyzes the brain's activity to decipher what people are seeing and thinking. Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ Huge thank you to LatticeFlow for sponsoring this episode. LatticeFlow - https://latticeflow.ai/ MLOps podcast #225 with Patrick Beukema, Head / Technical Lead of the Environmental AI, Applied Science Organization at AI2, Beyond AGI, Can AI Help Save the Planet? // Abstract AI will play a central role in solving some of our greatest environmental challenges. The technology that we need to solve these problems is in a nascent stage -- we are just getting started. For example, the combination of remote sensing (satellites) and high-performance AI operating at a global scale in real-time unlocks unprecedented avenues to new intelligence. MLOPs is often overlooked on AI teams, and typically there is a lot of friction in integrating software engineering best practices into the ML/AI workflow. However, performance ML/AI depends on extremely tight feedback loops from the user back to the model that enables high iteration velocity and ultimately continual improvement. We are making progress but environmental causes need your help. Join us fight for sustainability and conservation. // Bio Patrick is a machine learning engineer and scientist with a deep passion for leveraging artificial intelligence for social good. He currently leads the environmental AI team at the Allen Institute for Artificial Intelligence (AI2). His professional interests extend to enhancing scientific rigor in academia, where he is a strong advocate for the integration of professional software engineering practices to ensure reliability and reproducibility in academic research. Patrick holds a Ph.D. from the Center for Neuroscience at the University of Pittsburgh and the Center for the Neural Basis of Cognition at Carnegie Mellon University, where his research focused on neural plasticity and accelerated learning. He applied this expertise to develop state-of-the-art deep learning models for brain decoding of patient populations at a startup, later acquired by BlackRock. His earlier academic work spanned research on recurrent neural networks, causal inference, and ecology and biodiversity. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Variety of relevant papers/talks/links on Patrick's website: https://pbeukema.github.io/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Patrick on LinkedIn: https://www.linkedin.com/in/plbeukema/ Timestamps: [00:00] AI Quality Conference [01:29] Patrick's preferred coffee [02:00] Takeaways [04:14] Learning how to learn journey [07:04] Patrick's day to day [08:39] Environmental AI [11:07] Environmental AI models [14:35] Nature Inspires Scientific Advances [18:11] R&D [24:58] Iterative Feedback-Driven Development [26:37 - 28:07] LatticeFlow Ad [33:58] Balancing Metrics for Success [38:16] Model Retraining Pipeline [44:11] Series Models: Versatility [45:57] Edge Models Enhance Output [50:22] Custom Models for Specific Data [53:53] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Verena Weber believes that GenAI is going to transform the way we work and interact with devices. Her mission is to help companies prepare for this transformation. She has strong expertise in NLP and over 7 years of experience in Machine Learning. Huge thank you to @zilliz for sponsoring this episode. Zilliz - https://zilliz.com/ MLOps podcast #224 with Verena Weber, Generative AI Consultant at Verena Weber, GenAI in Production - Challenges and Trends. // Abstract The goal of this talk is to provide insights into challenges for Generative AI in production as well as trends aiming to solve some of these challenges. The challenges and trends Verena see are: Model size and moving towards mixture of experts architectures context window - new breakthroughs for context lengths from unimodality to multimodality, next step large action models? regulation in form of the EU AI Act Verena uses the differences between Gemini 1.0 and Gemini 1.5 to exemplify some of these trends. // Bio Verena leverages GenAI in natural language to elevate business competitiveness and navigate its transformative impact. Her varied experience in multiple roles and sectors underpins her ability to extract business value from AI, blending deep technical expertise with strong business acumen. Post-graduation, she consulted in Data Science at Deloitte and then advanced her skills in NLP, Deep Learning, and GenAI as a Research Scientist at Alexa team, Amazon. Passionate about gender diversity in tech, she mentors women to thrive in this field. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: verenaweber.de Sign up for Verena's newsletter: https://verenas-newsletter-63558b.beehiiv.com/ Zilliz - https://zilliz.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Verena on LinkedIn: https://www.linkedin.com/in/verena-weber-134178b9/ Timestamps: [00:00] AI Quality Conference [01:33] Verena's preferred coffee [02:15] Takeaways [06:33] Ski Person of Influence [11:31] Verena's background in the last 5-10 years [14:24] Tech Evolution: Rapid Transformation [18:13] Working at Amazon and key challenges [20:10] Research-inspired suggestions [22:21] AI Updates Impact Workflows [22:52] Alexa Query Distribution Analysis [24:06] Innovative Solutions for Alexa [25:27] Robust T5 Data Prompting [27:38] Audio Data Quality Challenges [28:21-29:28] Zilliz ad [29:28] Alexa data transcription and data cleaning
[35:38] Considering needs, costs, and complexity
[37:44] ChatGPt is not ideal for classification
[39:32] Comparison of model building using TF, IDF
[45:08] Struggle to boost diversity in conference speakers
[47:30] Creating safe environments helps underrepresented individuals participate
[48:29] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
MLOps Coffee Sessions Special episode with Databricks, Introducing DBRX: The Future of Language Models, fueled by our Premium Brand Partner, Databricks. DBRX is designed to be especially capable of a wide range of tasks and outperforms other open LLMs on standard benchmarks. It also promises to excel at code and math problems, areas where others have struggled. Our panel of experts will get into the technical nuances, potential applications, and implications of DBRx for businesses, developers, and the broader tech community. This session is a great opportunity to hear from insiders about how DBRX's capabilities can benefit you. // Bio Denny Lee - Co-host Denny Lee is a long-time Apache Spark™ and MLflow contributor, Delta Lake maintainer, and a Sr. Staff Developer Advocate at Databricks. A hands-on distributed systems and data sciences engineer with extensive experience developing internet-scale data platforms and predictive analytics systems. He has previously built enterprise DW/BI and big data systems at Microsoft, including Azure Cosmos DB, Project Isotope (HDInsight), and SQL Server. Davis Blalock Davis Blalock is a research scientist and the first employee at MosaicML. He previously worked at PocketSonics (acquired 2013) and completed his PhD at MIT, where he was advised by John Guttag. He received his M.S. from MIT and his B.S. from the University of Virginia. He is a Qualcomm Innovation Fellow, NSF Graduate Research Fellow, and Barry M. Goldwater Scholar. He is also the author of Davis Summarizes Papers, one of the most widely-read machine learning newsletters. Bandish Shah Bandish Shah is an Engineering Manager at MosaicML/Databricks, where he focuses on making generative AI training and inference efficient, fast, and accessible by bridging the gap between deep learning, large-scale distributed systems, and performance computing. Bandish has over a decade of experience building systems for machine learning and enterprise applications. Prior to MosaicML, Bandish held engineering and development roles at SambaNova Systems where he helped develop and ship the first RDU systems from the ground up, and Oracle where he worked as an ASIC engineer for SPARC-based enterprise servers. Abhi Venigalla Abhi is an NLP architect working on helping organizations build their own LLMs using Databricks. Joined as part of the MosaicML team and used to work as a researcher at Cerebras Systems. Ajay Saini Ajay is an engineering manager at Databricks leading the GenAI training platform team. He was one of the early engineers at MosaicML (acquired by Databricks) where he first helped build and launch Composer (an open source deep learning training framework) and afterwards led the development of the MosaicML training platform which enabled customers to train models (such as LLMs) from scratch on their own datasets at scale. Prior to MosaicML, Ajay was co-founder and CEO of Overfit, an online personal training startup (YC S20). Before that, Ajay worked on ML solutions for ransomware detection and data governance at Rubrik. Ajay has both a B.S. and MEng in computer science with a concentration in AI from MIT. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.databricks.com/ Databricks DBRX: https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ // Abstract Dive into the challenges of scaling AI models from Minimum Viable Product (MVP) to full production. The panel emphasizes the importance of continually updating knowledge and data, citing examples like teaching AI systems nuanced concepts and handling brand name translations.
User feedback's role in model training, alongside evaluation steps like human annotation and heuristic-based assessment, was highlighted.
The speakers stressed the necessity of tooling for user evaluation, version control, and regular performance updates. Insights on in-house and external tools for annotation and evaluation were shared, providing a comprehensive view of the complexities involved in scaling AI models. // Bio Alex Volkov - Moderator Alex Volkov is an AI Evangelist at Weights & Biases, celebrated for his expertise in clarifying the complexities of AI and advocating for its beneficial uses. He is the founder and host of ThursdAI, a weekly newsletter, and podcast that explores the latest in AI, its practical applications, open-source, and innovation. With a solid foundation as an AI startup founder and 20 years in full-stack software engineering, Alex offers a deep well of experience and insight into AI innovation. Eric Peter Product management leader and 2x founder with experience in enterprise products, data, and machine learning. Currently building tools for generative AI @Databricks. Donné Stevenson Focused on building AI-powered products that give companies the tools and expertise needed to harness to power of AI in their respective fields. Phillip Carter Phillip is on the product team at Honeycomb where he works on a bunch of different developer tooling things. He's an OpenTelemetry maintainer -- chances are if you've read the docs to learn how to use OTel, you've read his words. He's also Honeycomb's (accidental) prompt engineering expert by virtue of building and shipping products that use LLMs. In a past life, he worked on developer tools at Microsoft, helping bring the first cross-platform version of .NET into the world and grow to 5 million active developers. When not doing computer stuff, you'll find Phillip in the mountains riding a snowboard or backpacking in the Cascades. Andrew Hoh Andrew Hoh is the President and Co-Founder of LastMile AI. Previously, he was a Group PM Manager at Meta AI, driving product for their AI Platform. Previously, he was the Product Manager for the Machine Learning Infrastructure team at Airbnb and a founding team member of Azure Cosmos DB, Microsoft Azure's distributed NoSQL database. He graduated with a BA in Computer Science from Dartmouth College. A big thank you to our Premium Sponsors, @Databricks and @baseten for their generous support! // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
Shane Morris is now a Senior Executive Advisor at Devis. Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ Huge thank you to @WeightsBiases for sponsoring this episode. WandB Free Courses - https://wandb.ai/telidavies/ml-news/reports/Introducing-W-B-MLOps-Courses-Free-Course-Effective-MLOps-Model-Development--VmlldzozMDk2ODA2 MLOps podcast #223 with Shane Morris, Senior Executive Advisor of Devis, Data Engineering in the Federal Sector. // Abstract Let's focus on autonomous systems rather than automation, and then super-narrow it down to smaller, cheaper, and more accessible autonomous systems. // Bio Former music and entertainment data and software person somehow moves into defense and national security, with hilarious and predictable results. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Quality in Person Conference: https://www.aiqualityconference.com/ Website: https://shanemorris.sucks TikTok: https://www.tiktok.com/@shanemorrisdotsucks WandB Free Courses - https://wandb.ai/telidavies/ml-news/reports/Introducing-W-B-MLOps-Courses-Free-Course-Effective-MLOps-Model-Development--VmlldzozMDk2ODA2 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Shane on LinkedIn: https://www.linkedin.com/in/shanetollmanmorris/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Peter Guagenti is an accomplished business builder and entrepreneur with expertise in strategy, product development, marketing, sales, and operations. Peter has helped build multiple successful start-ups to exits, fueling high growth in each company along the way. He brings a broad perspective, deep problem-solving skills, the ability to drive innovation amongst teams, and a proven ability to convert strategy into action -- all backed up by a history of delivering results. Huge thank you to AWS for sponsoring this episode. AWS - https://aws.amazon.com/ MLOps podcast #222 with Peter Guagenti, President & CMO of Tabnine - What Business Stakeholders Want to See from the ML Teams. // Abstract Peter Guagenti shares his expertise in the tech industry, discussing topics from managing large-scale tech legacy applications and data experimentation to the evolution of the Internet. He returns to his history of building and transforming businesses, such as his work in the early 90s for People magazine's website and his current involvement in AI development for software companies. Guagenti discusses the use of predictive modeling in customer management and emphasizes the importance of re-architecting solutions to fit customer needs. He also delves deeper into the AI tools' effectiveness in software development and the value of maintaining privacy. Guagenti sees a bright future in AI democratization and shares his company's development of AI coding assistants. Discussing successful entrepreneurship, Guagenti highlights balancing technology and go-to-market strategies and the value of failing fast. // Bio Peter Guagenti is the President and Chief Marketing Officer at Tabnine. Guagenti is an accomplished business leader and entrepreneur with expertise in strategy, product development, marketing, sales, and operations. He most recently served as chief marketing officer at Cockroach Labs, and he previously held leadership positions at SingleStore, NGINX (acquired by F5 Networks), and Acquia (acquired by Vista Equity Partners). Guagenti also serves as an advisor to a number of visionary AI and data companies including DragonflyDB, Memgraph, and Treeverse. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Quality in Person Conference: https://www.aiqualityconference.com/ Measuring the impact of GitHub Copilot Survey: https://resources.github.com/learn/pathways/copilot/essentials/measuring-the-impact-of-github-copilot/ AWS Trainium and Inferentia: https://aws.amazon.com/machine-learning/trainium/ https://aws.amazon.com/machine-learning/inferentia/AI coding assistants: 8 features enterprises should seek: https://www.infoworld.com/article/3694900/ai-coding-assistants-8-features-enterprises-should-seek.htmlCareers at Tabnine: https://www.tabnine.com/careers --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Peter on LinkedIn: https://www.linkedin.com/in/peterguagenti/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Amritha Arun Babu Mysore has been an expert in the field of consumer electronics, software products, and online marketplaces for the past 15 years. She has experience developing supply chains from the ground up, delivering AI-based products to millions of users, and advocating for ethical AI across Amazon, Wayfair, Salesforce, and NetApp.
Abhik Choudhury is a Senior Analytics Managing Consultant and Data Scientist with 11 years of experience in designing and implementing scalable data solutions for organizations across various industries. Huge thank you to @latticeflow for sponsoring this episode. LatticeFlow - https://latticeflow.ai/ MLOps podcast #221 with Amritha Arun Babu Mysore, ML Product Leader at Klaviyo and Abhik Choudhury, Managing Consultant Analytics at IBM, MLOps - Design Thinking to Build ML Infra for ML and LLM Use Cases. // Abstract As machine learning (ML) and large language models (LLMs) continue permeating industries, robust ML infrastructure and operations (ML Ops) are crucial to deploying these AI systems successfully. This podcast discusses best practices for building reusable, scalable, and governable ML Ops architectures tailored to ML and LLM use cases. // Bio Amritha Arun Babu Mysore Amritha is an accomplished technology leader with over 12 years of experience spearheading product innovation and strategic initiatives at both large enterprises and rapid-growth startups. Leveraging her background in engineering, supply chain, and business, Amritha has led high-performing teams to deliver transformative solutions solving complex challenges. She has driven product road mapping, requirements analysis, system design, and launch execution for advanced platforms in domains like machine learning, logistics, and e-commerce. Abhik Choudhury Abhik is a Senior Analytics Managing Consultant and Data Scientist with 11 years of experience in designing and implementing scalable data solutions for organizations across various industries. Throughout his career, Abhik developed a strong understanding of AI/ML, Cloud computing, database management systems, data modeling, ETL processes, and Big Data Technologies. Abhik's expertise lies in leading cross-functional teams and collaborating with stakeholders at all levels to drive data-driven decision-making in longitudinal pharmacy and medical claims and wholesale drug distribution areas. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Quality in Person Conference in collaboration with Kolena: https://www.aiqualityconference.com/ LatticeFlow website: https://latticeflow.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abhik on LinkedIn: https://www.linkedin.com/in/abhik-choudhury-35450058 Connect with Amritha on LinkedIn: https://www.linkedin.com/in/amritha-arun-babu-a2273729/
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Demetrios Brinkmann is the founder of the MLOps Community. Brinkmann fell into the Machine Learning Operations world, and since, has interviewed the leading names around MLOps, Data Science, and Machine Learning. Huge thank you to Weights & Biases for sponsoring this episode. Weights & Biases - https://wandb.ai/site MLOps podcast #220 with our very own Founder of MLOps Community, Demetrios Brinkmann, Looking Back on 4 Years of the MLOps Community. // Abstract In this lively podcast episode, Mihail Eric hosts Demetrios Brinkmann, the founder of the MLOps Community, discussing its origin, structure, and challenges. Demetrios shares amusing tales of job hunting on LinkedIn and building the community despite lacking technical expertise, emphasizing the value of sharing and humor. They delve into the practicalities of hosting events, transitioning from self-funded to sponsorship-based, and tease upcoming activities with renowned speakers. Mihail and Demetrios explore job dynamics, the importance of sustained relationships, and diverse engagement methods like newsletters and volunteering. Demetrios reflects on his journey to Germany post-company closure, envisioning a global hub for AI learning, embodying the community's mission. // Bio At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building Lego houses with his daughter. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Quality in Person Conference in collaboration with Kolena: https://www.aiqualityconference.com/ Weights & Biases Free Course: https://wandb.ai/telidavies/ml-news/reports/Introducing-W-B-MLOps-Courses-Free-Course-Effective-MLOps-Model-Development--VmlldzozMDk2ODA2What Does Best in Class AI/ML Governance Look Like in Fin Services? // Charles Radclyffe // MLOps #2: https://youtu.be/l52sRMVPVk0 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/ Timestamps: [00:00] Demetrios preferred coffee and bizarre listening [01:44] The MLOps Community Brainchild [04:22] The MLOps Community today [07:15] AI Quality in Person Conference on June 25th! [08:42] Community Quality [10:00] Community Learnings and the Genesis [17:55] The 600 Mark [20:15] The Feedback form [22:52] Demetrios' Journey and Learnings [29:01] Building full tolerance [29:55] Weights & Biases Free Course Ad [34:52] Building community involvement for professional success and networking [38:52] Balance in Community Growth [43:56] Collection of volunteers [49:00] Events Challenges [53:28] The future of MLOps Community [59:40] "Caveman" lifestyle choice [1:00:45] Stronger Hallucinogen [1:02:30] Wrap up
Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/
Huge thank you to Databricks AI for sponsoring this episode. Databricks - http://databricks.com/
Bandish Shah is an Engineering Manager at MosaicML/Databricks, where he focuses on making generative AI training and inference efficient, fast, and accessible by bridging the gap between deep learning, large-scale distributed systems, and performance computing.
Davis Blalock is a Research Scientist and the first employee of Mosaic ML: a GenAI startup acquired for $1.3 billion by Databricks. MLOps podcast #219 with Databricks' Engineering Manager, Bandish Shah and Research Scientist Davis Blalock, The Art and Science of Training Large Language Models. // Abstract What's hard about language models at scale? Turns out...everything. MosaicML's Davis and Bandish share war stories and lessons learned from pushing the limits of LLM training and helping dozens of customers get LLMs into production. They cover what can go wrong at every level of the stack, how to make sure you're building the right solution, and some contrarian takes on the future of efficient models. // Bio Bandish Shah Bandish Shah is an Engineering Manager at MosaicML/Databricks, where he focuses on making generative AI training and inference efficient, fast, and accessible by bridging the gap between deep learning, large-scale distributed systems, and performance computing. Bandish has over a decade of experience building systems for machine learning and enterprise applications. Prior to MosaicML, Bandish held engineering and development roles at SambaNova Systems where he helped develop and ship the first RDU systems from the ground up, and Oracle where he worked as an ASIC engineer for SPARC-based enterprise servers. Davis Blalock Davis Blalock is a research scientist at MosaicML. He completed his PhD at MIT, advised by Professor John Guttag. His primary work is designing high-performance machine learning algorithms. He received his M.S. from MIT and his B.S. from the University of Virginia. He is a Qualcomm Innovation Fellow, NSF Graduate Research Fellow, and Barry M. Goldwater Scholar. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links
AI Quality In-person Conference: AI Quality in Person Conference: https://www.aiqualityconference.com/ Website: http://databricks.com/ Davis Summarizes Papers Newsletter signup link Davis' Newsletters: Learning to recognize spoken words from five unlabeled examples in under two seconds: https://arxiv.org/abs/1609.09196 Training on data at 5GB/s in a single thread: https://arxiv.org/abs/1808.02515 Nearest-neighbor searching through billions of images per second in one thread with no indexing: https://arxiv.org/abs/1706.10283 Multiplying matrices 10-100x faster than a matrix multiply (with some approximation error): https://arxiv.org/abs/2106.10860 Hidden Technical Debt in Machine Learning Systems: https://proceedings.neurips.cc/paper_files/paper/2015/file/86df7dcfd896fcaf2674f757a2463eba-Paper.pdf --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Davis on LinkedIn: https://www.linkedin.com/in/dblalock/ Connect with Bandish on LinkedIn: https://www.linkedin.com/in/bandish-shah/
// Abstract Diego, David, Ads, and Katharine, bring to light the risks, vulnerabilities, and evolving security landscape of machine learning as we venture into the AI-driven future. They underscore the importance of education in managing AI risks and the critical role privacy engineering plays in this narrative. They explore the legal and ethical implications of AI technologies, fostering a vital conversation on the balance between utility and privacy. // Bio Diego Oppenheimer - Moderator Diego Oppenheimer is a serial entrepreneur, product developer and investor with an extensive background in all things data. Currently, he is a Partner at Factory a venture fund specialized in AI investments as well as a co-founder at Guardrails AI. Previously he was an executive vice president at DataRobot, Founder and CEO at Algorithmia (acquired by DataRobot) and shipped some of Microsoft’s most used data analysis products including Excel, PowerBI and SQL Server. Diego is active in AI/ML communities as a founding member and strategic advisor for the AI Infrastructure Alliance and MLops.Community and works with leaders to define AI industry standards and best practices. Diego holds a Bachelor's degree in Information Systems and a Masters degree in Business Intelligence and Data Analytics from Carnegie Mellon University. Ads Dawson A mainly self-taught, driven, and motivated proficient application, network infrastructure & cyber security professional holding over eleven years experience from start-up to large-size enterprises leading the incident response process and specializing in extensive LLM/AI Security, Web Application Security and DevSecOps protecting REST API endpoints, large-scale microservice architectures in hybrid cloud environments, application source code as well as EDR, threat hunting, reverse engineering, and forensics. Ads have a passion for all things blue and red teams, be that offensive & API security, automation of detection & remediation (SOAR), or deep packet inspection for example. Ads is also a networking veteran and love a good PCAP to delve into. One of my favorite things at Defcon is hunting for PWNs at the "Wall of Sheep" village and inspecting malicious payloads and binaries. Katharine Jarmul Katharine Jarmul is a privacy activist and data scientist whose work and research focuses on privacy and security in data science workflows. She recently authored Practical Data Privacy for O'Reilly and works as a Principal Data Scientist at Thoughtworks. Katharine has held numerous leadership and independent contributor roles at large companies and startups in the US and Germany -- implementing data processing and machine learning systems with privacy and security built in and developing forward-looking, privacy-first data strategy. David Haber David has started and grown several technology companies. He developed safety-critical AI in the healthcare space and for autonomous flight. David has educated thousands of people and Fortune 500 companies on the topic of AI. Outside of work, he loves to spend time with his family and enjoys training for the next Ironman. A big thank you to our Premium Sponsors, @Databricks and @baseten for their generous support! // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
Frank Liu is the Director of Operations & ML Architect at Zilliz, where he serves as a maintainer for the Towhee open-source project. Jiang Chen is the Head of AI Platform and Ecosystem at Zilliz. Yujian Tang is a developer advocate at Zilliz. He has a background as a software engineer working on AutoML at Amazon. MLOps Coffee Sessions Special episode with Zilliz, Why Purpose-built Vector Databases Matter for Your Use Case, fueled by our Premium Brand Partner, Zilliz. Engineering deep-dive into the world of purpose-built databases optimized for vector data. In this live session, we explore why non-purpose-built databases fall short in handling vector data effectively and discuss real-world use cases demonstrating the transformative potential of purpose-built solutions. Whether you're a developer, data scientist, or database enthusiast, this virtual roundtable offers valuable insights into harnessing the full potential of vector data for your projects. // Bio Jiang Chen Frank Liu is Head of AI & ML at Zilliz, with over eight years of industry experience in machine learning and hardware engineering. Before joining Zilliz, Frank co-founded Orion Innovations, an IoT startup based in Shanghai, and worked as an ML Software Engineer at Yahoo in San Francisco. He presents at major industry events like the Open Source Summit and writes tech content for leading publications such as Towards Data Science and DZone. His passion for ML extends beyond the workplace; in his free time, he trains ML models and experiments with unique architectures. Frank holds MS and BS degrees in Electrical Engineering from Stanford University. Frank Liu Jiang Chen is the Head of AI Platform and Ecosystem at Zilliz. With years of experience in data infrastructures and information retrieval, Jiang previously served as a tech lead and product manager for Search Indexing at Google. Jiang holds a Master's degree in Computer Science from the University of Michigan, Ann Arbor. Yujian Tang Yujian Tang is a Developer Advocate at Zilliz. He has a background as a software engineer working on AutoML at Amazon. Yujian studied Computer Science, Statistics, and Neuroscience with research papers published to conferences including IEEE Big Data. He enjoys drinking bubble tea, spending time with family, and being near water. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://zilliz.com/ Neural Priming for Sample-Efficient Adaptation: https://arxiv.org/abs/2306.10191LIMA: Less Is More for Alignment: https://arxiv.org/abs/2305.11206ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT: https://arxiv.org/abs/2004.12832 Milvus Vector Database by Zilliz: https://zilliz.com/what-is-milvus --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Timestamps: [00:00] Demetrios' musical intro [04:36] Vector Databases vs. LLMs [07:51] Relevance Over Speed [12:55] Pipelines [16:19] Vector Databases Integration Benefits [26:42] Database Diversity Market [27:38] Milus vs. Pinecone [30:22] Vector DB for Training & Deployment [34:32] Future proof of AI applications [45:16] Data Size and Quality [48:53] ColBERT Model [54:25] Vector Data Consistency Best Practices [57:24] Wrap up
Huge thank you to LatticeFlow AI for sponsoring this episode. LatticeFlow AI - https://latticeflow.ai/.Dr. Petar Tsankov is a researcher and entrepreneur in the field of Computer Science and Artificial Intelligence. MLOps podcast #218 with Petar Tsankov, Co-Founder and CEO at LatticeFlow AI, A Decade of AI Safety and Trust. // Abstract // Bio Co-founder & CEO at LatticeFlow AI, building the world's first product enabling organizations to build performant, safe, and trustworthy AI systems. Before starting LatticeFlow AI, Petar was a senior researcher at ETH Zurich working on the security and reliability of modern systems, including deep learning models, smart contracts, and programmable networks. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://latticeflow.ai/ ERAN, the world's first scalable verifier for deep neural networks: https://github.com/eth-sri/eran VerX, the world's first fully automated verifier for smart contracts: https://verx.ch Securify, the first scalable security scanner for Ethereum smart contracts: https://securify.ch DeGuard, de-obfuscates Android binaries: http://apk-deguard.com SyNET, the first scalable network-wide configuration synthesis tool: https://synet.ethz.ch --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Petar on LinkedIn: https://www.linkedin.com/in/petartsankov/ Timestamps: [00:00] Petar's preferred coffee [00:29] Takeaways [03:15] Shout out to LatticeFlow for sponsoring this episode! [03:22] Please like, share, leave a review, and subscribe to our MLOps channels! [03:42] Expansion [05:16] Zurich ETH [07:06] AI Safety [09:24] Optimizing one metric, no fixed data sets [12:19] Trust life-changing issues [14:59] So much interest in GenAI [16:45] Explosion of GenAI Trust and Safety [21:14] Red Teaming [25:22] Trustworthy AI in Industry [27:43] DataOps Challenges [33:42] Trusting Third-Party Models [37:00] Testing Open Source Models [41:41] Specialized ML for Leasing [43:04] Regulation and Financial Incentives [45:30] Regulations Drive Innovation Balance [47:23] Regulations vs Certification: Voluntary Prove [52:24] Workflow Transparency: Trust & Efficiency [53:20] Engineers Balance Compliance Risks [54:53] Pushing Deep Learning Limits [57:31] Wrap up
Thank you to Zilliz our wonderful sponsors of this episode create some amazing stuff with Zilliz RAG - https://zilliz.com/vector-database-use-cases/llm-retrieval-augmented-generation
Sam Bean is a seasoned AI and machine learning expert, specializing in Large Language Models (LLMs) and search tech.
With a computer science background and a drive for innovation, Sam leads the team at Rewind AI in leveraging advanced tech to tackle complex challenges. MLOps podcast #217 with Sam Bean, Software Engineer (Applied AI) at Rewind.ai, The Real E2E RAG Stack. // Abstract What does a fully operational LLM + Search stack look like when you're running your own retrieval and inference infrastructure? What does the flywheel really mean for RAG applications? How do you maintain the quality of your responses? How do you prune/dedupe documents to maintain your document quality? // Bio Sam has been training, evaluating, and deploying production-grade inference solutions for language models for the past 2 years at You.com. Previous to that he built personalization algorithms at StockX. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://github.com/sam-h-bean/ REinforced Self Training (REST) - https://arxiv.org/pdf/2308.08998.pdf REST meets REACT - https://arxiv.org/pdf/2312.10003.pdf --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sam on LinkedIn: https://www.linkedin.com/in/samuel-h-bean/ Timestamps: [00:00] Sam's preferred coffee [00:11] Takeaways [03:52] A competitive coding pinball player [07:18] Sam's MLOps journey [10:33] Search Challenges with ML [15:04] Expensive evaluation [21:04] Labeling Parties Boost Data Quality [24:10] Zeno's Paradox of Motion [25:51] Sam's job at Rewind AI [29:35] Multimodal RAG [30:59 - 32:06] Zilliz Ad [32:07] University of Prague paper leak [36:38] Signals behind the scenes [39:28] Content Over Metadata Approach [43:22] Optionality around evaluation and search [48:35] Incremental Robustness Building [51:33] Solid Foundations for Success [53:42] Production RAGs [1:00:06] Thoughts on DSPy [1:05:40] Using DSPy in Production [1:08:26] Wrap up
Anass Bensrhir is the Associate Partner of McKinsey & Company Casablanca. Anu Arora is the Principal Data Engineering at McKinsey & Company.
Check out mckinsey.com/quantumblack MLOps podcast #214 with QuantumBlack AI by McKinsey's Principal Data Engineer, Anu Arora and Associate Partner, Anass Bensrhir, Managing Data for Effective GenAI Application brought to you by our Premium Brand Partner QuantumBlack AI by @McKinsey . // Abstract Generative AI is poised to bring impact across all industries and business functions across industries While many companies pilot GenAI, only a few have deployed GenAI use cases, e.g., retailers are producing videos to answer common customer questions using ChatGPT. A majority of organizations are facing challenges to industrialize and scale, with data being one of the biggest inhibitors. Organizations need to strengthen their data foundations given that among leading organizations, 72% noted managing data among the top challenges preventing them from scaling impact. Furthermore, leaders noticed that +31% of their staff's time is spent on non-value-added tasks due to poor data quality and availability issues. // Bio Anu Arora Data architect(~12 years) and have experience in Big data technologies, API development, building scalable data pipelines including DevOps and DataOps, and building GenAI solutions. Anass Bensrhir Anass Leads QuantumBlack in Africa, he specializes in the Financial sector and helps organizations deliver successful large Data transformation programs. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.mckinsey.com/capabilities/quantumblack/how-we-help-clients --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Anu on LinkedIn: https://uk.linkedin.com/in/anu-arora-072012 Connect with Anass on LinkedIn: https://www.linkedin.com/in/abensrhir/ Timestamps: [00:00] Anass and Anu's preferred coffee [00:35] Takeaways [04:02] Please like, share, leave a review, and subscribe to our MLOps channels! [04:09] Huge shout out to our sponsor QuantumBlack! [04:29] Anu's tech background [06:31] Anass tech background [07:28] The landscape of data [10:37] Dealing with unstructured data [15:51] Data lakes and ETL processes [22:19] Data Engineers' Heavy Workload [29:49] Data privacy and PII in the new LLMs paradigm [36:13] Balancing LLM Adoption Risk [44:06] Effective LMS Implementation Strategy [49:00] Decisions: Create or Wait [50:39] Wrap up
Alex Volkov serves as the AI Evangelist with Weights & Biases, Host of ThursdAI, Founder and CEO Targum and AI Consultant GPU POOR Def. not an owl.
MLOps podcast #215 with Alex Volkov, AI Evangelist at Weights & Biases, Becoming an AI Evangelist. // Abstract Follow Alex's journey into the world of AI, from being interested in running his first AI models to founding an AI startup, running a successful weekly AI news podcast & newsletter, and landing a job with @WeightsBiases . // Bio Alex Volkov is an AI Evangelist at Weights & Biases, celebrated for his expertise in clarifying the complexities of AI and advocating for its beneficial uses. He is the founder and host of ThursdAI, a weekly newsletter and podcast that explores the latest in AI, its practical applications, open-source and innovation. With a solid foundation as an AI startup founder and 20 years in full-stack software engineering, Alex offers a deep well of experience and insight into AI innovation. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Evaluation Survey: https://hq.yougot.us/primary/WebInterview/3AW6LW5D/Start Website: https://thursdai.news
Alex on X (+X spaces also are also there) - https://twitter.com/altryne/
ThursdAI podcast/newsletter - https://sub.thursdai.news
Denver local AI tinkerers meetup - https://denver-boulder.aitinkerers.org/
Weights & Biases Growth Team hack week review - https://www.youtube.com/watchInterview w/
Crew AI creator Joao Moura - https://sub.thursdai.news/p/jan14-sunday-special-deep-dives The Future of Search in the Era of Large Language Models // Saahil Jain // MLOps Podcast #150: https://youtu.be/hMoMvK89iogDSPy: Transforming Language Model Calls into Smart Pipelines // Omar Khattab // MLOps Podcast #194: https://youtu.be/NoaDWKHdkHg --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Alex on LinkedIn: https://www.linkedin.com/in/alex-volkov-/ Timestamps: [00:00] Alex's preferred beverage [00:17] Takeaways [03:19] Take the Evaluation Survey! [03:55] Alex's journey for the past 15 years [08:10] Career moves [15:15] Building communities [20:02] AI/MLOps Growth in COVID [27:23] Recent developments and insights [31:58 - 33:03] WandB Ad [33:54] Multimodal RAG and Lucid Dreaming [39:55] Evaluation Practices in MLOps [43:27] Evaluating AI models effectively [52:52] Embedding models and updates [56:14] AI model trade-offs [1:01:13] Optimizing LLM user experience [1:03:56] Perceived performance optimization [1:05:45] Agents' hype and reality [1:11:31] Exploring DSPy for evaluation [1:14:13] Wrap up
// Abstract
From startups achieving significant value with minor capabilities to AI revolutionizing sales calls and raising sales by 30%, we explore a series of interesting real-world use cases. Understanding the objectives and complexities of various industries, exploring the challenges of launching products, and highlighting the vital integration of the human touch with technology, this episode is a treasure trove of insights. // Bio Greg Kamradt - Moderator Greg has mentored thousands of developers and founders, empowering them to build AI-centric applications. By crafting tutorial-based content, Greg aims to guide everyone from seasoned builders to ambitious indie hackers. Some of his popular works: 'Introduction to LangChain Part 1, Part 2' (+145K views), and 'How To Question A Book' featuring Pinecone (+115K Views). Greg partners with companies during their product launches, feature enhancements, and funding rounds. His objective is to cultivate not just awareness, but also a practical understanding of how to optimally utilize a company's tools. He previously led Growth @ Salesforce for Sales & Service Clouds in addition to being early on at Digits, a FinTech Series-C company. Agnieszka Mikołajczyk-Bareła Senior AI Engineer@Chaptr working on LLMs. PhD, author of datasets, scientific papers, and publications with over 1800 citations, holding numerous scholarships and awards. Daily, she conducts her research on her grant "Detecting and overcoming bias in data with explainable artificial intelligence" Preludium, awarded by Polish National Centre. She is a co-organizer of PolEval2021 and PolEval 2022 tasks with punctuation prediction and restoration. She organizes and actively contributes to the scientific community in her free time: she managed and led the team during the HearAI project focused on modeling Sign Language. A former organizer and a team leader at the open-source project. As an ML Expert, she supports the project "Susana" designed to detect and read product expiry dates to help the Blind "see". Jason Liu Jason is a machine learning engineer and technical advisor. Arjun Kannan Arjun Kannan builds products, businesses, and teams. Currently building ResiDesk, bringing AI copilots to help real estate forecast renewals, reduce turnover, and hit their budget. Arjun built and led product and engineering functions at Climb Credit (serving 100k students, doubling loan growth for 3 years straight) and at BlackRock (creating $400mm in annual revenue), and helped build multiple startups and small companies before that. // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
Daniel Svonava is the Co-Founder of Superlinked. Daniel Svonava attended the Faculty of Informatics and Information Technologies, Slovak University of Technology. MLOps podcast #214 with Daniel Svonava, CEO & Co-founder at Superlinked, Information Retrieval & Relevance: Vector Embeddings for Semantic Search // Abstract In today's information-rich world, the ability to retrieve relevant information effectively is essential. This lecture explores the transformative power of vector embeddings, revolutionizing information retrieval by capturing semantic meaning and context. We'll delve into: - The fundamental concepts of vector embeddings and their role in semantic search - Techniques for creating meaningful vector representations of text and data - Algorithmic approaches for efficient vector similarity search and retrieval - Practical strategies for applying vector embeddings in information retrieval systems // Bio Daniel is an entrepreneurial technologist with a 20 year career starting with competitive programming and web development in highschool, algorithm research and Google & IBM Research internships during university, first entrepreneurial steps with his own computational photography startup and a 6 year tenure as a tech lead for ML infrastructure at YouTube Ads, where his ad performance forecasting engine powers the purchase of $10B of ads per year. Presently, Daniel is a co-founder of Superlinked.com - a ML infrastructure startup that makes it easier to build information-retrieval heavy systems - from Recommender Engines to Enterprise-focused LLM apps. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Daniel on LinkedIn: https://www.linkedin.com/in/svonava/?originalSubdomain=ch Timestamps: [00:00] Daniel's preferred coffee [00:13] Takeaways [04:59] Please like, share, leave a review, and subscribe to our MLOps channels! [05:22] Recommender system pivot insights [08:49] RaaS Challenges and solutions [10:23] Vector Compute vs Traditional Compute [13:20] String conversion challenges [17:18] Vector Computation in Recommender Systems [20:55] RAG system setup overview [26:00] ETL and Vector embeddings [31:04] Fine-tuning embedding models RAG [36:10] Flattening data for Vectors [37:18] Vector compute control insights [47:48] Vector Hub database comparison [51:22] Vector database partnership strategy [52:47] Vector computation in ML [55:43] Wrap up
Morgan McGuire has held a variety of roles in the past 13 years. In 2008, he completed a Research Internship at Queen Mary, University of London. Currently, he is the Head of Growth ML and Growth ML Engineer at Weights & Biases. Anish Shah has been working in the tech industry since 2015. In 2015, he was a Technical Support at Fox School of Business at Temple University. In 2021, he has been an MLOps Engineer - Growth and a Tier 2 Support Machine Learning Engineer at Weights & Biases. ______________________________________________ Large Language Models have taken the world by storm. But what are the real use cases? What are the challenges in productionizing them? In this event, you will hear from practitioners about how they are dealing with things such as cost optimization, latency requirements, trust of output, and debugging. You will also get the opportunity to join workshops that will teach you how to set up your use cases and skip over all the headaches. Join the AI in Production Conference on February 22 here: https://home.mlops.community/home/events/ai-in-production-2024-02-15 ______________________________________________ MLOps podcast #213 with Weights and Biases' Growth Director, Morgan McGuire and MLE, Anish Shah, Evaluating and Integrating ML Models brought to you by our Premium Brand Partner @WeightsBiases. // Abstract Anish Shah and Morgan McGuire share insights on their journey into ML, the exciting work they're doing at Weights and Biases, and their thoughts on MLOps. They discuss using large language models (LLMs) for translation, pre-written code, and internal support. They discuss the challenges of integrating LLMs into products, the need for real use cases, and maintaining credibility. They also touch on evaluating ML models collaboratively and the importance of continual improvement. They emphasize understanding retrieval and balancing novelty with precision. This episode provides a deep dive into Weights and Biases' work with LLMs and the future of ML evaluation in MLOps. It's a must-listen for anyone interested in LLMs and ML evaluation. // Bio Anish Shah Anish loves turning ML ideas into ML products. He started his career working with multiple Data Science teams within SAP, working with traditional ML, deep learning, and recommendation systems before landing at Weights & Biases. With the art of programming and a little magic, Anish crafts ML projects to help better serve our customers, turning “oh nos” to “a-ha”s! Morgan McGuire Morgan is a Growth Director and an ML Engineer at Weights & Biases. He has a background in NLP and previously worked at Facebook on the Safety team where he helped classify and flag potentially high-severity content for removal. // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI in Production Conference: https://home.mlops.community/home/events/ai-in-production-2024-02-15 Website: https://wandb.ai/ Prompt Templates the Song: https://www.youtube.com/watch?v=g6WT85gIsE8 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Morgan on LinkedIn: https://www.linkedin.com/in/morganmcg1/ Connect with Anish on LinkedIn: https://www.linkedin.com/in/anish-shah/
Alexandra Diem, PhD, has extensive experience in the field of AI, machine learning, and cloud analytics. Alexandra currently holds the position of Head of Cloud Analytics and MLOps at Gjensidige. Large Language Models have taken the world by storm. But what are the real use cases? What are the challenges in productionizing them? In this event, you will hear from practitioners about how they are dealing with things such as cost optimization, latency requirements, trust of output, and debugging. You will also get the opportunity to join workshops that will teach you how to set up your use cases and skip over all the headaches. Join the AI in Production Conference on February 22 here: https://home.mlops.community/home/events/ai-in-production-2024-02-15 _____________________________________________________________ MLOps podcast #212 with Alexandra Diem, Head of Cloud Analytics & MLOps at Gjensidige, Data Governance and AI. // Abstract This recent session featuring the incredibly talented Alexandra Diem delves into the challenges of generative AI in sensitive data environments, the emergence of specialized chatbots, and data governance. Balancing high-tech projects with those offering significant business value, using agile methods, is also discussed. Alexandra's journey from academia to being a consultant in Norway is truly inspiring. The discussion explores the function of enabling and R&D in tech roles, the shift towards self-serve solutions, and the integration of AI into existing workflows. Stimulating conversations about future-oriented technologies married with sound data science and industry practices make this session a must-listen for anyone interested in machine learning operations! // Bio Former academic turned data scientist with a passion for data mesh architectures. 🔬 Background in applied mathematics and statistics, adept at leveraging data-driven insights to solve complex problems. Experienced in diverse domains spanning the private and public sectors. 🧠 Made significant contributions to research in physiological modeling, successfully debunking a leading biomedical hypothesis on Alzheimer's disease during my PhD. Developed innovative approaches to quantify blood supply to the heart. 💡 Solution-oriented thinker with a track record of efficiently tackling challenging problems and adapting to novel scenarios. ⚙️ Expertise: Data Science | Mathematical Modeling | Statistical Analysis | Problem Solving In my spare time, you'll find me exploring the great outdoors—whether it's pedaling through scenic landscapes on a bike or riding down the slopes on a pair of skis. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI in Production Conference: https://home.mlops.community/home/events/ai-in-production-2024-02-15 Website: https://github.com/alexdiem Talk "DevOps revolutionised software engineering, it's time to revolutionise data" https://vimeo.com/861721829 from JavaZone 2023 Zilliz Cloud: https://zilliz.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Alexandra on LinkedIn: https://www.linkedin.com/in/dralexdiem/
Aayush Mudgal is a Senior Machine Learning Engineer at Pinterest, currently leading the efforts around Privacy-Aware Conversion Modeling. Large Language Models have taken the world by storm. But what are the real use cases? What are the challenges in productionizing them? In this event, you will hear from practitioners about how they are dealing with things such as cost optimization, latency requirements, trust of output, and debugging. You will also get the opportunity to join workshops that will teach you how to set up your use cases and skip over all the headaches. Join the AI in Production Conference on February 15 and 22 here: https://home.mlops.community/home/events/ai-in-production-2024-02-15 ________________________________________________________________________________________ MLOps podcast #211 with Aayush Mudgal, Senior Machine Learning Engineer at Pinterest, Ads Ranking Evolution at Pinterest. // Abstract Listen to the lessons from the journey of scaling ads ranking at Pinterest using innovative machine learning algorithms and innovation in the ML platform. Learn how they transitioned from traditional logistic regressions to deep learning-based transformer models, incorporating sequential signals, multi-task learning, and transfer learning. Discover the hurdles Pinterest overcame and the insights they gained in this talk, as Aayush shares the transformation of ads ranking at Pinterest and the lessons learned along the way. Discover how ML Platform evolution is crucial for algorithmic advancements. // Bio Aayush Mudgal is a Senior Machine Learning Engineer at Pinterest, currently leading the efforts around Privacy-Aware Conversion Modeling. He has a successful track record of starting and executing 0 to 1 projects, including conversion optimization, video ads ranking, landing page optimization, and evolving the ads ranking from GBDT to DNN stack. His expertise is in large-scale recommendation systems, personalization, and ads marketplaces. Before entering the industry, Aayush conducted research on intelligent tutoring systems, developing data-driven feedback to aid students in learning computer programming. He holds a Master's in Computer Science from Columbia University and a Bachelor of Technology in Computer Science from the Indian Institute of Technology Kanpur. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://www.youtube.com/watch?v=MZVIxtsGzBg https://www.youtube.com/watch?v=ffpPUr8Hg6U --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Aayush on LinkedIn: https://www.linkedin.com/in/aayushmudgal/
Large Language Models have taken the world by storm. But what are the real use cases? What are the challenges in productionizing them? In this event, you will hear from practitioners about how they are dealing with things such as cost optimization, latency requirements, trust of output, and debugging. You will also get the opportunity to join workshops that will teach you how to set up your use cases and skip over all the headaches.
Join the AI in Production Conference on February 15 and 22 here: https://home.mlops.community/home/events/ai-in-production-2024-02-15
________________________________________________________________________________________
Aparna Dhinakaran is the Co-Founder and Chief Product Officer at Arize AI, a pioneer and early leader in machine learning (ML) observability.
MLOps podcast #210 with Aparna Dhinakaran, Co-Founder and Chief Product Officer of Arize AI, LLM Evaluation with Arize AI's Aparna Dhinakaran.
// Abstract
Dive into the complexities of Language Model (LLM) evaluation, the role of the Phoenix evaluations library, and the importance of highly customized evaluations in software application. The discourse delves into the nuances of fine-tuning in AI, the debate between the use of open-source versus private models, and the urgency of getting models into production for early identification of bottlenecks. Then examine the relevance of retrieved information, output legitimacy, and the operational advantages of Phoenix in supporting LLM evaluations.
// Bio
Aparna Dhinakaran is the Co-Founder and Chief Product Officer at Arize AI, a pioneer and early leader in AI observability and LLM evaluation. A frequent speaker at top conferences and thought leader in the space, Dhinakaran is a Forbes 30 Under 30 honoree. Before Arize, Dhinakaran was an ML engineer and leader at Uber, Apple, and TubeMogul (acquired by Adobe). During her time at Uber, she built several core ML Infrastructure platforms, including Michelangelo. She has a bachelor’s from Berkeley's Electrical Engineering and Computer Science program, where she published research with Berkeley's AI Research group. She is on a leave of absence from the Computer Vision Ph.D. program at Cornell University.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Arize-Phoenix: https://phoenix.arize.com/
Phoenix LLM task eval library: https://docs.arize.com/phoenix/llm-evals/running-pre-tested-evals
Aparna's recent piece on LLM evaluation: https://arize.com/blog-course/llm-evaluation-the-definitive-guide/
Thread on the difference between model and task LLM evals: https://twitter.com/aparnadhinak/status/1752763354320404488
Research thread on why numeric score evals are broken: https://twitter.com/aparnadhinak/status/1748368364395721128
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Aparna on LinkedIn: https://www.linkedin.com/in/aparnadhinakaran/
Matt Bleifer is a Group Product Manager at Tecton, where he focuses on the core product experience such as building, testing, and productionizing feature pipelines as scale. Michael Eastham works as a Chief Architect at Tecton, which is a Business Intelligence (BI) Software company with an estimated 100 employees Large Language Models have taken the world by storm. But what are the real use cases? What are the challenges in productionizing them? In this event, you will hear from practitioners about how they are dealing with things such as cost optimization, latency requirements, trust of output, and debugging. You will also get the opportunity to join workshops that will teach you how to set up your use cases and skip over all the headaches. Join the AI in Production Conference on February 15 and 22 here: https://home.mlops.community/home/events/ai-in-production-2024-02-15 ________________________________________________________________________________________ MLOps podcast #209 with Tecton's Group Product Manager, Matt Bleifer and Chief Architect, Mike Eastham, Powering MLOps: The Story of Tecton's Rift brought to us by our Premium Brand Partner, @tecton8241 . // Abstract Explore the intricacies of feature platforms and their integration in the data realm. Compare traditional predictive machine learning with the integration of Linguistic Model Systems into software applications. Get a glimpse of Rift, a product enhancing data processing with smooth compatibility with various technologies. Join in on the journey of developing Rift, and making Tecton user-friendly, and enjoy Matt's insights and contributions. Wrap it up with lighthearted talks on future collaborations, music, and a touch of nostalgia. // Bio Matt Bleifer Matt Bleifer is a Group Product Manager and an early employee at Tecton. He focuses on core product experiences such as building, testing, and productionizing feature pipelines at scale. Before joining Tecton, he was a Product Manager for Machine Learning at both Twitter and Workday, totaling nearly a decade of working on machine learning platforms. Matt has a Bachelor’s Degree in Computer Science from California Polytechnic State University, San Luis Obispo. Michael Eastham Michael Eastham is the Chief at Tecton. Previously, he was a software engineer at Google, working on Web Search. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.tecton.ai/Rift Article: https://www.tecton.ai/blog/unlocking-real-time-ai-for-everyone-with-tecton/ Rift: https://resources.tecton.ai/riftBig Data is Dead blog: https://motherduck.com/blog/big-data-is-dead/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Matt on LinkedIn: https://www.linkedin.com/in/mattbleifer/Connect with Mike on LinkedIn: https://www.linkedin.com/in/mikeeastham/ Timestamps: [00:00] AI in Production Conference [02:13] Matt & Mike's preferred coffee [02:37] Takeaways [04:50] Matt & Mike's Tecton titles [06:49] Matt's background in tech [09:49] Mike's background in tech [12:53] Tecton refresher [18:23] Feature store to Feature platform [21:06] Current evolution of Tecton [24:41] The understatement [26:12] Duck DB Con [27:54] Rift [30:10] Kafka Flink [33:36] What is large in aggregations? [38:09] Big Data is Dead! [41:14] Principles of creating Rift [45:54] The battle between Simplicity and Flexibility [47:28] Is he serious? Segment [50:54] Can you get any more hype Segment [57:10] What are you excited about? [1:02:45] Wrap up
Join our virtual conference 'AI in Production'
Transform faster. Innovate smarter. Anticipate the future. At QuantumBlack, we unlock the power of artificial intelligence (AI) to help organizations reinvent themselves from the ground up—and accelerate sustainable and inclusive growth.
MLOps Coffee Sessions Special episode with QuantumBlack, AI by McKinsey, GenAI Buy vs Build, Commercial vs Open Source, fueled by our Premium Brand Partner, QuantumBlack, AI by McKinsey. // Abstract Do you build or buy? Check the QuantumBlack team discussing the different sides of buying vs building your own GenAI solution. Let's look at the trade-offs companies need to make - including some of the considerations of using black box solutions that do not provide transparency on what data sources were used. Whether you are a business leader or a developer exploring the space of GenAI, this talk provides you with valuable insights to prepare you for how you can be more informed and prepared for navigating this fast-moving space. // Bio Ilona Logvinova Ilona Logvinova is the Head of Innovation for McKinsey Legal, working across the legal department to identify, lead, and implement cross-cutting and impactful innovation initiatives, covering legal technologies and reimagination of the profession initiatives. At McKinsey Ilona is also Managing Counsel for McKinsey Digital, working closely with emerging technologies across use cases and industries. Mohamed Abusaid Am Mohamed, a tech enthusiast, hacker, avid traveler, and foodie all rolled into one individual. Built his first website when he was 9 and fell in love with computers and the internet ever since. Graduated with a computer science from university although dabbled in electrical, electronic, and network engineering before that. When he's not reading up on the latest tech conversations and products on Hacker News, Mohamed spends his time traveling to new destinations and exploring their cuisine and culture. Mohamed works with different companies helping them tackle challenges in developing, deploying, and scaling their analytics to reach its potential. Some topics he's enthusiastic about include MLOps, DataOps, GenerativeAI, Product thinking, and building cross-functional teams to deliver user-first products. Nayur Khan Nayur is a partner within McKinsey and part of the QuantumBlack, AI by McKinsey leadership team. He predominantly focuses on helping organizations build capabilities to industrialize and scale artificial intelligence (AI), including the newer Generative AI. He helps companies navigate innovations, technologies, processes, and digital skills as needed to run at scale. He is a keynote speaker and is recognized in the DataIQ 100 - a list of the top 100 influential people in data. Nayur also leads the firm’s diversity and inclusion efforts within QuantumBlack to promote a more equitable environment for all. He speaks with organizations on the importance of diversity and diverse team building—especially when working with data and building AI. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ilona on LinkedIn: www.linkedin.com/in/ilonalogvinova Connect with Mo on LinkedIn: https://www.linkedin.com/in/mabusaid/ Connect with Nayur on LinkedIn: https://www.linkedin.com/in/nayur/
Jon Cooke is the owner/founder of Dataception a Data, Analytics, and Data Product company, and the creator of the Data Product Pyramid, an adaptive Data Product operating model.
MLOps podcast #208 with Jon Cooke, CTO of Dataception, Micro Graph Transformer - Specialist Small Language Models Using Graphs to Accelerate Data Product Eco-systems. // Abstract Specialist deconstructed Encoder/Decoder Transformers along with data product management and tech to vastly accelerate prototyping, building, and deploying business-facing data products at high speed and low cost. // Bio Jon is a 20-year veteran in Data, Analytics, and AI and is a Data product specialist. After many times seeing the massive time, friction, failures, and costs typically associated with data and analytics initiatives, Jon founded Dataception. Its mission is to use tech to eliminate the data grunt and work together with data product management and AI to build and iterate sophisticated, business-facing analytics in real-time in front of the business. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI in Production Conference: https://home.mlops.community/public/events/ai-in-production-2024-02-15 Website: www.dataception.com https://www.linkedin.com/events/generativeai-dataproductsandbus7114951387100184576/theater/ https://www.linkedin.com/events/12thevalueofadataproductmanagem7110920848416366594/comments/ https://www.linkedin.com/events/howtoactuallyusedataproductstod7113570339535638528/theater/Building Better Data Teams // Leanne Fitzpatrick // Coffee Sessions #113: https://www.youtube.com/watch?v=JxVS3-4wyKc --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jon on LinkedIn: https://www.linkedin.com/in/jon-cooke-096bb0/
Timestamps: [00:00] AI in Production Conference teaser [02:12] Jon's preferred coffee [02:24] Takeaways [03:48] Please like, share, and subscribe to our MLOps channels! [04:02] Backpacking, traveling, and almost cast for Lord of the Rings [06:40] Jon's tech background [11:07] Dataception [15:05] Data Challenges: Delays & Causes [16:46] Data Virtualization for Agility [19:47] Large Company Change Challenges [21:28] Sales Tools Migration Challenges [24:44] Data and ML Integration [28:13] Data Roles Evolution [32:20] Tech for Prototyping Acceleration [35:22] LLM Enables Natural Language Data Analytics [36:36] Ensuring Reliable AI Information [38:20] Proxy Routing and Intelligent Agents [42:41] Human API for Data [46:49] Engineer Success with Growth [48:15] Tech CEO Balancing Act [53:59] Iterative Development for Product-Market Fit [56:16] Wrap up
Jake Watson is the writer of thedataplatform.substack.com and Principal Data Engineer at The Oakland Group.
MLOps podcast #207 with Jake Watson, Principal Data Engineer at The Oakland Group, How Data Platforms Affect ML & AI. // Abstract I’ve always told my clients and colleagues that traditional rule-based software is difficult, but software containing Artificial Intelligence (AI) and/or Machine Learning (ML)* is even more difficult, sometimes impossible. Why is this the case? Well, software is difficult because it’s like flying a plane while building it at the same time, but because AI and ML make rules on the fly based on various factors like training data, it’s like trying to build a plane in flight, but some parts of the plane will be designed by a machine, and you have little idea what that is going to look like till the machine finishes. This double goes for more cutting-edge AI models like GPT, where only the creators of the software have a vague idea of what it will output. This makes software with AI / ML more of a scientific experiment than engineering, which is going to make your project manager lose their mind when you have little idea how long a task is going to take. But what will make everyone’s lives easier is having solid data foundations to work from. Learn to walk before running. // Bio Jake has been working in data as an Analyst, Engineer, and/or Architect for over 10 years. Started as an analyst in the UK National Health Service converting spreadsheets to databases tracking surgical instruments. Then continued as an analyst at a consultancy (Capita) reporting on employee engagement in the NHS and dozens of UK Universities. There Jake moved reporting from Excel and Access to SQL Server, Python with frontend websites in d3.js. At Oakland Group, a data consultancy, Jake worked as a Cloud Engineer, Data Engineer, Tech Lead, and Architect depending on the project for dozens of clients both big and small (mostly big). Jake has also developed and productionised ML solutions as well in the NLP and classification space. Jake has experience in building Data Platforms in Azure, AWS, and GCP (though mostly in Azure and AWS) using Infrastructure as Code and DevOps/DataOps/MLOps. In the last year, Jake has been writing articles and newsletters for my blog, including a guide on how to build a data platform: https://thedataplatform.substack.com/p/how-to-build-a-data-platform // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://thedataplatform.substack.com/ How Data Platform Foundations Impact AI and ML Applications blog: https://thedataplatform.substack.com/p/issue-29-how-data-platform-foundations AI in Production Conference: https://home.mlops.community/public/events/ai-in-production-2024-02-15 How to Build a Data Platform blog: https://thedataplatform.substack.com/p/how-to-build-a-data-platform --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jake on LinkedIn: https://www.linkedin.com/in/jake-watson-data/ Timestamps: [00:00] Jake's preferred coffee [00:26] AI in Production Conference teaser [02:38] Takeaways [04:00] Please like, share, and subscribe to our MLOps channels! [04:17] Data Engineer's Crucial Role [05:44] Jake's background [06:44] Data Platform Foundations blog [10:34] Data mesh organizational side of things [17:58] Importance of data modeling [20:13] Dealing with the sprawl [22:03] Data quality [23:59] Data hierarchy on building a platform [29:34] ML Platform Team Structure [31:47] Don't reinvent the wheel [34:04] Data pipelines synergy [37:31] Wrap up
Yujian is working as a Developer Advocate at Zilliz, where they develop and write tutorials for proof of concepts for large language model applications. They also give talks on vector databases, LLM Apps, semantic search, and tangential spaces.
MLOps podcast #206 with Yujian Tang, Developer Advocate at Zilliz, RAG Has Been Oversimplified, brought to us by our Premium Brand Partner, Zilliz // Abstract In the world of development, Retrieval Augmented Generation (RAG) has often been oversimplified. Despite the industry's push, the practical application of RAG reveals complexities beyond its apparent simplicity. This talk delves into the nuanced challenges and considerations developers encounter when working with RAG, providing a candid exploration of the intricacies often overlooked in the broader narrative. // Bio Yujian Tang is a Developer Advocate at Zilliz. He has a background as a software engineer working on AutoML at Amazon. Yujian studied Computer Science, Statistics, and Neuroscience with research papers published to conferences including IEEE Big Data. He enjoys drinking bubble tea, spending time with family, and being near water. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: zilliz.com --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Yujian on LinkedIn: linkedin.com/in/yujiantang Timestamps: [00:00] Yujian's preferred coffee [00:17] Takeaways [02:42] Please like, share, and subscribe to our MLOps channels! [02:55] The hero of the LLM space [05:42] Embeddings into Vector databases [09:15] What is large and what is small LLM consensus [10:10] QA Bot behind the scenes [13:59] Fun fact getting more context [17:05] RAGs eliminate the ability of LLMs to hallucinate [18:50] Critical part of the rag stack [19:57] Building citations [20:48] Difference between context and relevance [26:11] Missing prompt tooling [27:46] Similarity search [29:54] RAG Optimization [33:03] Interacting with LLMs and tradeoffs [35:22] RAGs not suited for [39:33] Fashion App [42:43] Multimodel Rags vs LLM RAGs [44:18] Multimodel use cases [46:50] Video citations [47:31] Wrap up
Jonathan Frankle works as Chief Scientist (Neural Networks) at MosaicML (recently acquired by Databricks), a startup dedicated to making it easy and cost-effective for anyone to train large-scale, state-of-the-art neural networks. He leads the research team. MLOps podcast #205 with Jonathan Frankle, Chief Scientist (Neural Networks) at Databricks, The Myth of AI Breakthroughs, co-hosted by Denny Lee, brought to us by our Premium Brand Partner, Databricks. // Abstract Jonathan takes us behind the scenes of the rigorous work they undertake to test new knowledge in AI and to create effective and efficient model training tools. With a knack for cutting through the hype, Jonathan focuses on the realities and usefulness of AI and its application. We delve into issues such as face recognition systems, the 'lottery ticket hypothesis,' and robust decision-making protocols for training models. Our discussion extends into Jonathan's interesting move into the world of law as an adjunct professor, the need for healthy scientific discourse, his experience with GPUs, and the amusing claim of a revolutionary algorithm called Qstar. // Bio Jonathan Frankle is Chief Scientist (Neural Networks) at Databricks, where he leads the research team toward the goal of developing more efficient algorithms for training neural networks. He arrived via Databricks’ $1.3B acquisition of MosaicML as part of the founding team. He recently completed his PhD at MIT, where he empirically studied deep learning with Prof. Michael Carbin, specifically the properties of sparse networks that allow them to train effectively (his "Lottery Ticket Hypothesis" - ICLR 2019 Best Paper). In addition to his technical work, he is actively involved in policymaking around challenges related to machine learning. He earned his BSE and MSE in computer science at Princeton and has previously spent time at Google Brain and Facebook AI Research as an intern and Georgetown Law as an Adjunct Professor of Law. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.jfrankle.com Facial recognition: perpetuallineup.orgThe Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networksby Jonathan Frankle and Michael Carbin paper: https://arxiv.org/abs/1803.03635 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Denny on LinkedIn: https://linkedin.com/in/dennyglee Connect with Jonathan on LinkedIn: https://www.linkedin.com/in/jfrankle/ Timestamps: [00:00] Jonathan's preferred coffee [01:16] Takeaways [07:19] LM Avalanche Panel Surprise [10:07] Adjunct Professor of Law [12:59] Low facial recognition accuracy [14:22] Automated decision making human in the loop argument [16:09] Control vs. Outsourcing Concerns [18:02] perpetuallineup.org [23:41] Face Recognition Challenges [26:18] The lottery ticket hypothesis [29:20] Mosaic Role: Model Expertise [31:40] Expertise Integration in Training [38:19] SLURM opinions [41:30] GPU Affinity [45:04] Breakthroughs with QStar [49:52] Deciphering the noise advice [53:07] Real Conversations [55:47] How to cut through the noise [1:00:12] Research Iterations and Timelines [1:02:30] User Interests, Model Limits [1:06:18] Debugability [1:08:00] Wrap up
Patrick Barker is the Founder / CTO of Kentauros AI.
Farhood Etaati is a Software Engineer at Yektanet. MLOps podcast #204 with Patrick Barker, CTO of Kentauros AI and Farhood Etaati, MLOps/Platform Team Lead at AIMedic, MLOps at the Crossroads. // Abstract MLOps is at a crossroads. The ever-increasing excitement for LLMs' ability to solve some interesting real-world problems has made many people interested in applying these models in new applications which comes with its own challenges, that have upstarted the term "LLMLOps". But how much of those challenges are not a newer representation of what older-gen ML models had to deal with in the production, and the question arises whether developing "new" specialized tools to address these applications actually provides any substantial value for the sustainability of the field in general. Tools are coming and going at a rate that makes many technical people skeptical of adopting newer tools. What can we do as a community to alleviate these issues? Why OSS MLOps is lacking behind and how VC money is contributing to that? // Bio Farhood Etaati MLOps engineer at AIMedic. Studied EE at Uni of Tehran, started out as a data scientist, and pivoted to software engineering. Currently working on on-premise MLOps platform development suitable for Iran's infrastructure. Patrick Barker When Patrick is not occupied with building his AI company, he enjoys spending time with his wonderful kids or exploring the hills of Boulder. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Websites: https://github.com/pbarker
--------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Patrick on LinkedIn: https://www.linkedin.com/in/patrickbarkerco/ Connect with Farhood on LinkedIn: www.linkedin.com/in/farhood-etaati Timestamps: [00:00] Farhood's and Patrick's preferred coffee [01:13] Takeaways [04:00] Please like, share, and subscribe to our MLOps channels! [05:26] Strong feelings [10:21] MLOps vs DevOps Challenges [13:44] Medical setting, ML tools, NLP, model building [16:23] MLOps vs Data Engineering [20:45] MLOps Boosts LLM Development [23:54] Longtail Use Cases [31:00] Tech Roles Distinctions [34:42] Did He Say That? [37:04] Fine-tuning AI Models [38:57] ML 2.0 Advancements Explained [41:11] Generative AI in MLOps [45:04] ML Reproducibility Challenges [48:03] Wrap up
Aleksa Gordić is an ex-Google DeepMind / Microsoft ML engineer currently working on non-English LLMs at OrtusAI, open-sourcing Meta's NLLB (no language left behind) project and YugoGPT.
MLOps podcast #203 with Aleksa Gordić, Founder of OrtusAI, Pioneering AI Models for Regional Languages. // Abstract Dive deep into Aleksa's work with the YugoGPT, a language model serving Serbian, Croatian, Bosnian, and Montenegrin dialects - emphasizing the need for multilingual AI developments. Explore the unique language dynamics in the Balkans and Eastern Europe, the potential business opportunities around multilingual models, and the challenges in deploying large language models. Aleksa shares his experience with vision and image models, his collaborations with key tech players, and his use of advanced technologies. Hear about Aleksa Gordić's journey of being active and visible in the tech community and his insights into the world of machine learning and AI. Prepare to have your thinking challenged and horizons widened as we converse about the intriguing and complex world of MLOps. // Bio Working on non-English LLMs at OrtusAI, open-sourcing Meta's NLLB (no language left behind) project. Worked at DeepMind on the Flamingo project as a research engineer. Worked at Microsoft on the HoloLens 2 project & next-gen mixed reality glasses. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://gordicaleksa.com/ https://github.com/gordicaleksa - I build stuff :) https://discord.com/invite/peBrCpheKE - active AI Discord server (~6000) I bring the best AI researchers in the world to give talks (James Betker DALL-E 3 author, Tri Dao (Flash Attention), etc.) https://gordicaleksa.medium.com/how-i-got-a-job-at-deepmind-as-a-research-engineer-without-a-machine-learning-degree-1a45f2a781de - how I landed a job at DeepMind (and a couple more potentially interesting writings) Aleksa Gordić The AI Epiphany Youtube Channel: https://www.youtube.com/channel/UCj8shE7aIn4Yawwbo2FceCQ/videos W&B AI Academy: http://wandb.me/mlops_com_llm_course --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Aleksa on LinkedIn: https://www.linkedin.com/in/aleksagordic/ Timestamps: [00:00] Aleksa's preferred coffee [00:17] Takeaways [02:51] Humming the GPU's [06:23] Built Chrome extension for communicating with videos [08:04] Rig Doubles Throughput Time [09:32] Vector databases advise [10:38] Learning from experts, connecting, and gathering insights. [13:47] Zero to Hero for MLOps [15:37] Serendipitous moments [17:52] Depth Over Breaking News [19:50] Trust in GPT Content [22:22] Exam Challenges and AI [26:53] YugoGPT [31:41] WandB Ad [33:33] Linguistic Mysteries [34:52] No Language Left Behind project (NLLB project) [36:53] YugoGPT Development Overview [37:49] NLLB vs YugoGPT [39:35] Yugo GPT parameters [41:16] Opportunities for unsupported languages [43:08] Diffusion model [44:39] Generative AI with image generation models [47:45] AI Challenges and Excitement [50:32] Challenges in different alphabet characters [52:10] Need a co-founder [56:05] Career transition and entrepreneurial mindset [1:00:20] Big Tech salary misconceptions [1:03:02] Inspiring wrap up
Prof. Dr. Hannes Mühleisen is a creator of the DuckDB database management system and Co-founder and CEO of DuckDB Labs. Jordan is co-founder and chief duck-herder at MotherDuck, a startup building a serverless analytics platform based on DuckDB.
MLOps podcast #202 with Hannes Mühleisen, Co-Founder & CEO of DuckDB Labs and Jordan Tigani, Chief Duck-Herder at MotherDuck, Small Data, Big Impact: The Story Behind DuckDB. // Abstract Navigate the intricacies of data management with Jordan Tagani and Hannes Mühleisen, the creative geniuses behind DuckDB and MotherDuck. This deep dive unravels the game-changing principles behind DuckDB's creation, tackling the prevailing wisdom to passionately fill the gap for smaller data set management. Let's also discover MotherDuck's unique focus on providing an unprecedented developer experience and its innovative edge in visualization and data delivery. This episode is teeming with enlightening discussions about managing community feedback, funding, and future possibilities that should not be missed for any tech enthusiasts and data management practitioners. // Bio Hannes Mühleisen Prof. Dr. Hannes Mühleisen is a creator of the DuckDB database management system and Co-founder and CEO of DuckDB Labs, a consulting company providing services around DuckDB. Hannes is also Professor of Data Engineering at Radboud Universiteit Nijmegen. His' main interest is analytical data management systems. Jordan Tigani Jordan is co-founder and chief duck-herder at MotherDuck, a startup building a serverless analytics platform based on DuckDB. He spent a decade working on Google BigQuery, as a founding engineer, book author, engineering leader, and product leader. More recently, as SingleStore’s Chief Product Officer, Jordan helped them build a cloud-native SaaS business. Jordan has also worked at Microsoft Research, the Windows Kernel team, and at a handful of star-crossed startups. His biggest claim to fame is predicting world cup matches using machine learning with a better record than Paul the Octopus. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Websites: https://duckdb.org/ https://motherduck.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Hannes on LinkedIn: https://www.linkedin.com/in/hfmuehleisen/ Connect with Jordan on LinkedIn: https://www.linkedin.com/in/jordantigani/ Timestamps: [00:00] Hannes and Jordan's preferred coffee [01:30] Takeaways [03:43] Swaggers in the house! [07:13] Duck DB's inception [09:38] Jordan's background [12:28] Simplify Developer Experience [17:54] Big Data Shift [26:01] Creation of MotherDuck [30:58] Duck DB and MotherDuck Partnership [31:57] Incentive Alignment Concerns [37:46] Building an incredible developer experience [43:38] User Testing Lab [47:18] Setting a higher standard [49:22] The moments before the moment [52:18] Gathering feedback and talking to the community [54:30] MotherDuck Features [1:00:19] Cloud Innovation for MotherDuck [1:02:41] ML Engineers and DuckDB [1:08:03] Wrap up
Paco Nathan is the Managing Partner at Derwen, Inc., and author of Latent Space, along with other books, plus popular videos and tutorials about machine learning, natural language, graph technologies, and related topics.
MLOps podcast #201 with Paco Nathan, Managing Partner at Derwen, Inc., Language, Graphs, and AI in Industry. // Abstract Let's talk about key findings from these conferences, specifically summarizing teams that have ROI on machine learning in production: what are the things in common they're doing, and what are the most important caveats they urge other teams to consider when getting started? Because these key takeaways aren't found in the current AI news cycle. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Conference: https://aiconference.com/ K1st World: https://www.k1st.world/ Corunna Innovation Summit: https://corunna.dataspartan.com/ "Cloud Computing on Amazon AWS EC2" UC Berkeley EECS guest lecture (2009) https://vimeo.com/manage/videos/3616394 "Hardware - Software - Process: Data Science in a Post-Moore’s Law World" https://www.nvidia.com/en-us/ai-data-science/resources/hardware-software-process-book/ “LLMs in Production: Learning from Experience” by Waleed Kadous @ Anyscale https://www.youtube.com/watch?v=xa7k9MUeIdk "Supercharging Industrial Operations with Problem-Solving GenAI & Domain Knowledge" by Christopher Nguyen @ Aitomatic https://www.k1st.world/2023-program/supercharging-industrial-operations-with-problem-solving-genai-domain-knowledge “The Next Million AI Systems” by Mark Huang @ Gradient: https://www.youtube.com/watch?v=lA0Npe4PqFw "AI in a Box" by Useful Sensors https://usefulsensors.com/#products "Opportunities in AI - 2023" by Andrew Ng https://www.youtube.com/watch?v=5p248yoa3oE "Advancing the Marine Industry Through the Harmony of Fishermen Knowledge and Al" by Akinori Kasai @ Furuno https://www.k1st.world/2023-program/advancing-the-marine-industry-through-the-harmony-of-fishermen-knowledge-and-al Macy conferences (1941-1960) https://en.wikipedia.org/wiki/Macy_conferences https://www.asc-cybernetics.org/foundations/history/MacySummary.htm https://press.uchicago.edu/ucp/books/book/distributed/C/bo23348570.html second-order cybernetics https://pangaro.com/designconversation/wp-content/uploads/dubberly-pangaro-chk-journal-2015.pdf https://en.wikipedia.org/wiki/Second-order_cybernetics Project Cybersyn https://jacobin.com/2015/04/allende-chile-beer-medina-cybersyn/ https://thereader.mitpress.mit.edu/project-cybersyn-chiles-radical-experiment-in-cybernetic-socialism/ https://99percentinvisible.org/episode/project-cybersyn/ https://medium.com/@rjog/project-cybersyn-an-early-attempt-at-iot-governance-and-how-we-can-apply-its-learnings-5164be850413 https://www.sustema.com/post/project-cybersyn-how-a-chilean-government-almost-controlled-the-economy-from-a-control-room https://transform-social.org/en/texts/cybersyn/ Humberto Maturana, Francisco Varela: Autopoeisis "De Maquinas y Seres Vivos" "Everything said is said by an observer" https://proyectos.yura.website/wp-content/uploads/2021/06/de_maquinas_y_seres_vivos_-_maturana.pdf https://en.wikipedia.org/wiki/Autopoiesis_and_Cognition:_The_Realization_of_the_Living Fernando Flores (led Project Cybersyn, imprisoned, later worked with Prof. Terry Winograd @ Stanford, the grad advisor for what became Google) https://lorenabarba.com/gallery/prof-barba-gave-keynote-at-pycon-2016/ https://conversationsforaction.com/fernando-flores "Navigating the Risk Landscape: A Deep Dive into Generative AI" by Ben Lorica and Andrew Burt https://thedataexchange.media/mitigating-generative-ai-risks/ "SpanMarker" by Tom Aarsen @ Hugging Face https://tomaarsen.github.io/SpanMarkerNER/ Examples of "the math catching up with the machine learning": Guy Van den Broeck @ UCLA https://web.cs.ucla.edu/~guyvdb/talks/ Charles Martin @ Calculations Consulting https://weightwatcher.ai/
Mihail Eric is an engineer, researcher, and educator who has helped start teams at innovative organizations such as Amazon Alexa and RideOS.
Mihail is a cofounder of Storia AI where they build an AI-powered creative assistant for fast and delightful image and video generation.
MLOps podcast #200 with Mihail Eric, Co-founder of Storia AI, Founding, Funding, and the Future of MLOps. // Abstract Demetrios and Mihail journey deep into the significance of human sentiment in an increasingly AI-driven era, the perils and promises of conversational AI, and the evolution and impact of image generation models. Delve into the world of MLOps versus LLMOps, offering clarifying perspectives on how the core concerns and technology persist, even amidst an evolving tech landscape with new buzzwords making waves. Mihail generously provides an inside look at his AI tool and its wide range of applications across various industries, offering insights on interesting niche-specific verticals and unexpected use cases. // Bio Mihail is a co-CEO of Storia AI, an early-stage startup building an AI-powered creative assistant for video production. He has over a decade of experience researching and engineering AI systems at scale. Previously he built the first deep-learning dialogue systems at the Stanford NLP group. He was also a founding member of Amazon Alexa’s first special projects team where he built the organization’s earliest large language models. Mihail is a serial entrepreneur who previously founded Confetti AI, a machine-learning education company that he led until its acquisition in 2022. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: www.mihaileric.com
https://www.storia.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/ Timestamps: [00:00] Spotify playlist [02:08] How to live a longer healthier life [03:43] Absurd sweater collection [06:03] The catch-up episode [08:10] MLOps versus LLMOps [13:27] AI apps mixing visuals and text [16:00] Founder dating [16:54] Stable diffusion and difficulties with Mid Journey [23:12] Stripe developer experience [25:04] APIs and Model Providers [27:33] Host stable diffusion on AWS [34:07] AI Creativity: Prompt Experimentation [35:45] AI Challenges and Solutions [39:51] AI Hype Frustration [44:31] AI Impact on Hollywood [48:11] AI Impact on Filmmaking [51:48] Generalizable Tool for Verticals [52:49] MLOps versus LLMOps [56:42] Wrap up
Nathan Ryan Frank is the Machine Learning Operations and platform Director of Grainger. Former Astrophysicist turned data scientist and machine learning engineer with a proven history of delivering results into production across a wide variety of domains while leading projects with international, cross-functional teams. MLOps podcast #199 with Nathan Ryan Frank, Director, Machine Learning Platform & Operations at WW Grainger, Challenges Operationalizing Machine Learning (And Some Solutions). // Abstract This talk details some common challenges and pitfalls when attempting to operationalize machine learning systems and discusses some simple solutions. We dive into the machine learning development workflow and cover topics such as team dynamics, communication issues between roles that don't share a common language, and approaching MLOps from an SRE/DevOps perspective. Similarly, the talk highlights some challenges unique to operationalizing machine learning, drawing distinctions where necessary to highlight a large amount of similarity. Finally, the talk offers some simple and practical guidance for those new to MLOps who want to understand where to start and how to adopt best practices in an evolving field. // Bio Nathan Frank is currently the Director of Machine Learning Platform and Operations at Grainger where he is building a team to support the Technology Group's expanding machine learning efforts. Prior to joining Grainger, Nathan led machine learning engineering efforts at Strong Analytics, a boutique data science and machine learning consulting firm, as well as machine learning platform and development teams at Stats Perform, a leader in sports data and technology. Nathan holds bachelor's and master's degrees in Astrophysics from UC - Santa Cruz and UNC-Chapel Hill, respectively. When not building machine learning systems, Nathan spends as much time as possible with his favorite person in the world, his wife, as well as their four kids and two dogs, and enjoys getting outside to hike or garden and baking bread. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://nrfrank.github.io/ Bisi: https://bisi.gitbook.io/bisi/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Nathan on LinkedIn: https://www.linkedin.com/in/nrfrank Timestamps: [00:00] Nathan's preferred coffee [00:40] Takeaways [02:00] Please leave a review in our comment sections! Please like, share, and subscribe to our MLOps channels! [03:00] Telescope for gamma-ray burst [07:31] Transition into ML [11:23] Stats-heavy US sports commentary [14:25] Building machine learning systems approach [20:02] ML Workflow Must-Haves [26:50] Love for tests [33:10] Test Writing Importance [34:37] Bridging Stakeholder Language Gap [43:04] Shared Language, Team Collaboration [47:28] Rapid fire questions [51:20] Wrap up
Nick Hasty is the Director of Product, Discovery & Machine Learning at Giphy, an animated-gif search engine that allows users to search, share, and discover GIFs.
MLOps podcast #198, Inferring Creativity. // Abstract Generative AI models have captured our imaginations with their ability to produce new "creative" works such as visually striking images, poems, stories, etc, and their outputs often rival or excel what most humans can do. I believe that these developments should make us re-think the nature of creativity itself, and through identifying parallels and differences between generative models and the human brain we can establish a framework to talk about creativity, and its relationship to intelligence, that should hold up against future revelations in ML and neuroscience. // Bio Nick Hasty is a technologist & entrepreneur with a background in the creative arts. He was the founding engineer for GIPHY, where he’s worked for the last 10 years and now leads ML/AI product initiatives. He also servers as a consultant helping early-stage startups scale their product and engineering teams. Before GIPHY, he worked with arts+cultural organizations such as Rhizome.org and the Alan Lomax archives. He got his graduate degree from NYU’s ITP program. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: http://jnhasty.com/ Previous talks: https://engineering.giphy.com/giphy2vec-natural-language-processing-giphy/ https://changelog.com/practicalai/38 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Nick on LinkedIn: https://www.linkedin.com/in/nickhasty/ Timestamps: [00:00] Nick's preferred coffee [00:15] Takeaways [06:58] Nick's background in ML [12:15] Nick's GIPHY journey [17:39] Nick's Success Factors [20:50] The trajectory of AI [28:09] Identifying as a product engineer [32:42] Evaluate LLMs vs. Traditional Models [35:03] AI Product: Intuition vs Data [38:53] Giphy AI Product Development [45:25] Startups and Venture Assistance [52:30] AI Funding Landscape Shift [54:00] Wrap up
MLOps podcast #197 with Niels Bantilan, Chief Machine Learning Engineer at Union, The Role of Infrastructure in ML Leveraging Open Source brought to us by Union. // Abstract When we start out building and deploying models in a new organization, life is simple: all I need to do is grab some data, iterate on a model that fits the data well and performs reasonably well on some held-out test set. Then, if you’re fortunate enough to get to the point where you want to deploy it, it’s fairly straightforward to wrap it in an app framework and host it on a cloud server. However, once you get past this stage, you’re likely to find yourself needing: More scalable data processing framework Experiment tracking for models Heavier duty CPU/GPU hardware Versioning tools to link models, data, code, and resource requirements Monitoring tools for tracking data and model quality There’s a rich ecosystem of open-source tools that solves each of these problems and more: but how do you unify all of them together into a single view? This is where orchestration tools like Flyte can help. Flyte not only allows you to compose data and ML pipelines, but it also serves as “infrastructure as code” so that you can leverage the open-source ecosystem and unify purpose-built tools for different parts of the ML lifecycle on a single platform. ML systems are not just models: they are the models, data, and infrastructure combined. // Bio Niels is the Chief Machine Learning Engineer at Union.ai, and core maintainer of Flyte, an open-source workflow orchestration tool, author of UnionML, an MLOps framework for machine learning microservices, and creator of Pandera, a statistical typing and data testing tool for scientific data containers. His mission is to help data science and machine learning practitioners be more productive. He has a Masters in Public Health with a specialization in sociomedical science and public health informatics, and prior to that a background in developmental biology and immunology. His research interests include reinforcement learning, AutoML, creative machine learning, and fairness, accountability, and transparency in automated systems. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://github.com/cosmicBboy, https://union.ai/Flyte: https://flyte.org/ MLOps vs ML Orchestration // Ketan Umare // MLOps Podcast #183 - https://youtu.be/k2QRNJXyzFg --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Niels on LinkedIn: https://www.linkedin.com/in/nbantilan/ Timestamps: [00:00] Niels' preferred coffee [00:17] Takeaways [03:45] Shout out to our Premium Brand Partner, Union! [04:30] Pandera [08:12] Creating a company [14:22] Injecting ML for Data [17:30] ML for Infrastructure Optimization [22:17] AI Implementation Challenges [24:25] Generative DevOps movement [28:27] Pushing Limits: Code Responsibility [29:46] Orchestration in OpenAI's Dev Day [34:27] MLOps Stack: Layers & Challenges [42:45] Mature Companies Embrace Kubernetes [45:29] Horizon Challenges [47:24] Flexible Integration for Resources [49:10] MLOps Reproducibility Challenges [53:14] MLOps Maturity Spectrum [57:48] First-Class Citizens in Design [1:00:16] Delegating for Efficient Collaboration [1:04:55] Wrap up
Laurel Orr is a Principal Engineer at Numbers Station, a startup that applies Foundation Model technology to the enterprise data stack. Venky Orr is SVP of Product & Engineering.
MLOps podcast #196 with Numbers Station's Venky Ganti SVP, Product & Engineering and Principal Engineer, Laurel Orr, LLMs in Focus: From One-Size Fits All to Verticalized Solutions. // Abstract Dive into the realm of large language models (LLMs) as we explore the merits and limitations of 'one-size fits all' LLMs, and their role in data analytics. Through customer stories, we showcase real-world applications and contrast general LLMs with verticalized, enterprise-centric models. We address the significance of ownership structures, with a focus on open-source vs proprietary impacts on transparency and trustworthiness. Delving into the NSQL foundation models, we emphasize the importance of diverse, quality training data, especially with enterprise challenges. Lastly, we speculate on the future of LLMs, highlighting hosting solutions and the evolution towards specialized challenges. // Bio Laurel Orr Laurel Orr is a Principal Engineer at Numbers Station, a startup that applies Foundation Model technology to the enterprise data stack. Her research interests include how to use FMs to solve classically hard data-wrangling tasks and how to put FM technology into deployment. Before Numbers Station, Laurel was a postdoc at Stanford advised by Chris Re as part of the Hazy Research Labs working in the intersection of AI and data management. She graduated with a PhD in database systems from the University of Washington. Venky Orr Venky brings over two decades of experience in software engineering and technical leadership to Numbers Station as SVP of Product & Engineering. Most recently, he served as General Manager leading several initiatives on query understanding and commerce in the ads product area at Google. Before that, he was CEO and co-founder of Mesh Dynamics, the API test automation company, which was acquired by Google in 2021. Prior to Mesh Dynamics, Venky was CTO and co-founder of Alation, the enterprise data catalog company, where he led technology and helped create the new data catalog product category. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.numbersstation.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Laurel on LinkedIn: https://www.linkedin.com/in/laurel-orr/ Connect with Venky on LinkedIn: https://www.linkedin.com/in/venky-ganti-2679a2/ Timestamps: [00:00] Venky's and Laurel's preferred coffee [00:36] Takeaways [03:15] Please like, share, and subscribe to our MLOps channels! [04:38] Venky's background [07:47] Laurel's at background [09:38] Data wrangling [13:45] Sequel query [19:25] One size-fits-all LLMs vs Verticalized and Specific LLMs [23:42] Model Choice Trade-offs [30:18] NSQL Foundational Models [37:26] LLM Trends in 12 Months [40:09] Data recipes being democratized [45:16] Claude and 100,000 Context [48:02] Exploring Varieties of LLMs [50:02] AI Gateway [51:07] Text-to-SQL Model Evaluation [54:00] Wrap up
MLOps Coffee Sessions Special episode with Weights & Biases, Model Management in a Regulated Environment, fueled by our Premium Brand Partner, Weights & Biases. // Abstract Step into the fascinating world of Language Model Management (LLMs) in a Regulated Environment! Join us for an enlightening chat where we'll explore the intricacies of managing models within highly regulated settings, focusing on compliance and effective strategies. This is your opportunity to be part of a dynamic conversation that delves into the challenges and best practices of Model Management in Regulated Environments. Secure your spot today and stay tuned for an enriching dialogue on navigating the complexities of navigating the regulated terrain. Don't miss out on the chance to broaden your understanding and connect with peers in the field! // Bio Darek Kłeczek Darek Kłeczek is a Machine Learning Engineer at Weights & Biases, where he leads the W&B education program. Previously, he applied machine learning across supply chain, manufacturing, legal, and commercial use cases. He also worked on operationalizing machine learning at P&G. Darek contributed the first Polish versions of BERT and GPT language models and is a Kaggle Competitions Grandmaster. Mark Huang Mark is a co-founder and Chief Architect at Gradient, a platform that helps companies build custom AI applications by making it extremely easy to fine-tune foundational models and deploy them into production. Previously, he was a tech lead in machine learning teams at Splunk and Box, developing and deploying production systems for streaming analytics, personalization, and forecasting. Prior to his career in software development, he was an algorithmic trader at quantitative hedge funds where he also harnessed large-scale data to generate trading signals for billion-dollar asset portfolios. Oliver Chipperfield Oliver Chipperfield is a Senior Data Scientist and Team Lead at M-KOPA, where he utilizes his expertise in machine learning and data-driven innovation. At M-KOPA since October 2021, Oliver leads a diverse tech team, making improvements in credit loss forecasting and fraud detection. His career spans multiple industries, where he has applied his extensive knowledge in Python, Spark, R, SQL, and Excel. He also specialized in the building and design of production ML systems, experimentation, and Bayesian statistics. Michelle Marie Conway As an Irish woman who relocated to London after completing her university studies in Dublin, Michelle spent the past 12 years carving out a career in the data and tech industry. With a keen eye for detail and a passion for innovation, She has consistently leveraged my expertise to drive growth and deliver results for the companies she worked for. As a dynamic and driven professional, Michelle is always looking for new challenges and opportunities to learn and grow, and she's excited to see what the future holds in this exciting and ever-evolving industry. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Fine-Tuning LLMs: Best Practices and When to Go Small // Mark Kim-Huang // MLOps Meetup #124 - https://youtu.be/1WSUfWojoe0 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Darek on LinkedIn: https://www.linkedin.com/in/kleczek/ Connect with Mark on LinkedIn: https://www.linkedin.com/in/markhng525/ Connect with Oliver on LinkedIn: https://www.linkedin.com/in/oliver-chipperfield/ Connect with Michelle on LinkedIn: https://www.linkedin.com/in/michelle-conway-40337432
MLOps podcast #195 with Varun Mohan, CEO of Codeium, Building the Future of AI in Software Development brought to us by QuantumBlack. // Abstract This brief overview traces the evolution of Exafunction and Codeium, highlighting the strategic transition. It explores the inception of Codeium's key features, offering insights into the thoughtful design process. This emphasizes the company's forward-looking approach to preparing for a rapidly advancing technological landscape. Additionally, it touches upon developing essential MLOps systems, showcasing the commitment to maintaining rigor and efficiency in the face of evolving challenges. // Bio Varun Mohan developed a knack for programming in high school where he actively participated in various competitions. This passion for programming was shared with his now co-founder, with whom he frequently competed. Their common interest in programming and competition led them to attend MIT together, where they undertook more programming challenges. After college, they ventured into the Bay Area where they continued to compete and further cultivate their programming abilities. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Websites: codeium.com, https://exafunction.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Varun on Twitter: https://www.linkedin.com/in/varunkmohan/ Timestamps: [00:00] Varun's preferred coffee [00:15] Takeaways [02:50] Please like, share, and subscribe to our MLOps channels! [03:05] QuantumBlack ad by Nayur Khan [05:51] Varun's background in tech [10:55] Language Models Advancement [14:17] GPU scarce world [18:23] Vision and Pain Points [19:18] Fine-tuning Challenges in NLP [21:04] ML and AI Caution [21:49] MLOps: App vs Infra [23:53] Data Engineering Abstraction Evolution [26:12] Codeium and Scaling Discussion [31:59] API, Cloud, Computation [34:20] Codeium scaling [35:11] Reserved GPUs, companies self-hosting products [38:00] Open-source code Codeium training [40:03] Protecting IP Licenses [41:32] ML Challenges: Data, Bias, Security [44:37] Evaluating code [48:29] Getting values from Codeium [49:49] Exafunction ML AI Production [52:17] AWS Creation [53:58] Feature flags and MA AI lifecycle [56:34] Coding problem [58:40] New software architectures [1:03:28] Wrap up
// Abstract Explore the transformative role of AI in EdTech, discussing its potential to enhance learning experiences and personalize education. The panelists share insights on AI use cases, challenges in AI integration, and strategies for building a differentiated business model in the evolving AI landscape. The discussion looks ahead at how the latest wave of GenAI is set to shape the future of education. Join us to understand the exciting prospects and challenges of AI in EdTech. Moderator: Paul van der Boor // Bio Klinton Bicknell Klinton Bicknell is the Head of AI @duolingo . He works at the intersection of artificial intelligence and cognitive science. His research has been published in venues including ACL, PNAS, NAACL, Psychological Science, EDM, CogSci, and Cognition, and covered in the Financial Times, BBC, and Forbes. Prior to Duolingo, he was an assistant professor at Northwestern University. Bill Salak Bill Salak has more than 20 years of experience overseeing large-scale development projects and more than 24 years of experience in web application architecture and development. Bill founded and served as CTO of multiple Internet and web development companies, leading technology projects for companies including Age of Learning, AOL, Educational Testing Systems, Film LA, Hasbro, HBO, Highlights for Children, NBC-Universal, and the U.S. Army. Bill currently serves as the CTO of @Brainly-app , the leading learning platform worldwide with the most extensive Knowledge Base for all school subjects and grades. Yeva Hyusyan Yeva Hyusyan is the Co-Founder and CEO of @Sololearn , the most engaging platform for learning how to code. Prior to co-founding SoloLearn, Yeva established a startup accelerator for mobile games, consumer apps, and ag-tech solutions. In a previous role, she implemented programs for the World Bank and the US Government in business and education. Later, she served as a General Manager at Microsoft, where she led sales, developer ecosystem development, and strategic partnerships. Yeva holds an MBA in Corporate Strategy from Maastricht School of Management in the Netherlands, an MS in International Economics from Yerevan State University in Armenia, and completed the Executive Program at Stanford University's Graduate School of Business. // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
Join our conference: https://home.mlops.community/public/events/llms-in-production-part-iii-2023-10-03
MLOps Coffee Sessions Special episode with Tecton, Get your ML Application Into Production, sponsored by Tecton. // Abstract Getting an ML application into production is more difficult than most teams expect—but with the right preparation, it can be done efficiently! Join us for this exclusive roundtable, where 4 machine learning experts from Tecton will discuss some of the most common challenges and best practices to avoid them. With over 35 years of combined experience in MLOps at companies like AWS, Google, Lyft, and Uber, and 15 years of experience at Tecton spent helping customers like FanDuel, Plaid, and HelloFresh getting ML models into production, the presenters will share how factors like organizational structure, use cases, tech stack, and more, can create different types of bottlenecks. They’ll also share best practices and lessons learned throughout their careers on how to overcome these challenges. // Bio Kevin Stumpf Kevin co-founded Tecton where he leads a world-class engineering team that is building a next-generation feature store for operational Machine Learning. Kevin and his co-founders built deep expertise in operational ML platforms while at Uber, where they created the Michelangelo platform that enabled Uber to scale from 0 to 1000's of ML-driven applications in just a few years. Prior to Uber, Kevin founded Dispatcher, with the vision to build the Uber for long-haul trucking. Kevin holds an MBA from Stanford University and a Bachelor's Degree in Computer and Management Sciences from the University of Hagen. Outside of work, Kevin is a passionate long-distance endurance athlete. Derek Salama Derek is currently a Senior Product Manager at Tecton, where he is responsible for security, collaboration experience, and Feature Platform infrastructure. Prior to Tecton, Derek worked at Google and Lyft across both ML infrastructure and ML applications. Eddie Esquivel Eddie Esquivel is a Solutions Architect at Tecton, where he helps customers implement feature stores as part of their stack for operational ML. Prior to Tecton, Eddie was a Solutions Architect at AWS. He holds a Bachelor’s Degree in Computer Science & Engineering from the University of California, Los Angeles. Isaac Cameron Isaac Cameron is a Consulting Architect at Tecton. Prior to Tecton, he was a Principal Solutions Architect at Slalom Build, focusing on data and machine learning, where he built his own feature platform for a large U.S. airline and has enabled many organizations to build intelligent products leveraging operational ML. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Kevin on LinkedIn: https://www.linkedin.com/in/kevinstumpf/ Connect with Derek on LinkedIn: https://www.linkedin.com/in/dereksalama/ Connect with Eddie on LinkedIn: https://www.linkedin.com/in/eddie-esquivel-2016/ Connect with Isaac on LinkedIn: https://www.linkedin.com/in/isaaccameron/ Timestamps: [00:00] Introduction to Kevin Stumpf, Derek Salama, Eddie Esquivel, and Isaac Cameron [02:48] Challenges of traditional classical ML into production [10:21] Infrastructure cost [16:50] Bridging Business and Tech [19:23] ML Infrastructure Essentials [29:38] Integrated Batch and Stream [35:12] Scaling AI from Zero [36:23] Stacks red flags [45:53] Tecton: Features Quality Monitoring [49:06] Building Recommender System Tools [53:19] Quantify business value in ML [54:40] Wrap up
MLOps podcast #194 with Omar Khattab, PhD Candidate at Stanford, DSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines. // Abstract The ML community is rapidly exploring techniques for prompting language models (LMs) and for stacking them into pipelines that solve complex tasks. Unfortunately, existing LM pipelines are typically implemented using hard-coded "prompt templates", i.e. lengthy strings discovered via trial and error. Toward a more systematic approach for developing and optimizing LM pipelines, we introduce DSPy, a programming model that abstracts LM pipelines as text transformation graphs, i.e. imperative computational graphs where LMs are invoked through declarative modules. DSPy modules are parameterized, meaning they can learn (by creating and collecting demonstrations) how to apply compositions of prompting, finetuning, augmentation, and reasoning techniques. We design a compiler that will optimize any DSPy pipeline to maximize a given metric. We conduct two case studies, showing that succinct DSPy programs can express and optimize sophisticated LM pipelines that reason about math word problems, tackle multi-hop retrieval, answer complex questions, and control agent loops. Within minutes of compiling, a few lines of DSPy allow GPT-3.5 and llama2-13b-chat to self-bootstrap pipelines that outperform standard few-shot prompting and pipelines with expert-created demonstrations. On top of that, DSPy programs compiled to open and relatively small LMs like 770M-parameter T5 and llama2-13b-chat are competitive with approaches that rely on expert-written prompt chains for proprietary GPT-3.5. DSPy is available as open source at https://github.com/stanfordnlp/dspy // Bio Omar Khattab is a PhD candidate at Stanford and an Apple PhD Scholar in AI/ML. He builds retrieval models as well as retrieval-based NLP systems, which can leverage large text collections to craft knowledgeable responses efficiently and transparently. Omar is the author of the ColBERT retrieval model, which has been central to the development of the field of neural retrieval, and author of several of its derivate NLP systems like ColBERT-QA and Baleen. His recent work includes the DSPy framework for solving advanced tasks with language models (LMs) and retrieval models (RMs). // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://omarkhattab.com/ DSPy: https://github.com/stanfordnlp/dspy --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Omar on Twitter: https://twitter.com/lateinteraction Timestamps: [00:00] Omar's preferred coffee [00:26] Takeaways [06:40] Weight & Biases Ad [09:00] Omar's tech background [13:35] Evolution of RAG [16:33] Complex retrievals [21:32] Vector Encoding for Databases [23:50] BERT vs New Models [28:00] Resilient Pipelines: Design Principles [33:37] MLOps Workflow Challenges [36:15] Guiding LLMs for Tasks [37:40] Large Language Models: Usage and Costs [41:32] DSPy Breakdown [51:05] AI Compliance Roundtable [55:40] Fine-Tuning Frustrations and Solutions [57:27] Fine-Tuning Challenges in ML [1:00:55] Versatile GPT-3 in Agents [1:03:53] AI Focus: DSP and Retrieval [1:04:55] Commercialization plans [1:05:27] Wrap up
// Abstract Martian is focused on building a model router to dynamically route every prompt to the best LLM for the highest performance and lowest cost. Corti, the Al Co-Pilot for health care uses Al to improve patient care, demonstrating the potential of Al in healthcare and medical decision-making. They recently raised $60M, with Prosus being one of the lead investors. Transforms is pioneering in synthetic entertainment, showing how Al can transform the way we create and consume media. Moderator: Paul van der Boor // Speakers Sandeep Bakshi Head of Investments, Europe @prosusgroup3707 Shriyash Upadhyay Founder @Martian Lars Maaløe Co-Founder & CTO at Corti | Adj. Assoc. Professor of Machine Learning @ Corti Pietro Gagliano President & Founder @Transitional Forms // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
MLOps podcast #193 with Pierre Salvy, Head of Engineering at Cambrium, LLM in Material Production co-hosted by Stephen Batifol. // Abstract Delve into the world of proteins, genetic engineering, and the intersection of AI and biotech. Pierre explains how his company is using advanced models to design proteins with specific properties, even creating a vegan collagen for cosmetics. By harnessing the potential of AI, they aim to revolutionize sustainability, uncovering a future of lab-grown meats, molecular cheese, and less harmful plastics, confronting regulatory barriers and decoding the syntax and grammar of proteins. // Bio Head of Engineering at Cambrium, a biotech company utilising genAI to design sustainable protein biomaterials for the future. Pierre spent the last decade researching ways to make computers calculate better biological systems. This is a critical step to engineering more sustainable ways to make the products we use every day, which is their mission at Cambrium. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: cambrium.bio --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stephen-batifol/ Connect with Pierre on LinkedIn: https://www.linkedin.com/in/psalvy/ Timestamps: [00:00] Pierre's preferred coffee [00:10] Takeaways [05:10] Please like, share, and subscribe to our MLOps channels! [05:25] Weights and Biases ad [07:52] Ski story [09:54] Pierre's career trajectory [13:35] From employee #2 to hiring a team [14:42] From employee #2 to head of engineering [15:50] Uncomfortable things to say essential for growth and effectiveness [18:27] From biotech to engineering [21:10] LLMs at Cambrium [24:26] Slackbot [25:43] Quick and Easy Solutions [26:47] Products created at Cambrium [31:56] Impact of EU Regulation on Cambrium [35:39] 2nd Biotech Winter [36:35] Cost of error vs service not working [38:00] Protein Synthesis and Mutations [40:03] Large-Scale System Engineering Challenges [43:28] Expensive Factors in Experiments [44:39] LLMs vs Protein Models [47:03] Protein Design with LLMs [49:43] Eco-Friendly Product Vision [53:28] Space glue [54:00] Wrap up
// Abstract A product-minded engineering perspective on UX/design patterns, product evaluation, and building with AI. // Bio Charles Frye Charles teaches people how to build ML applications. After doing research in psychopharmacology and neurobiology, he pivoted to artificial neural networks and completed a PhD at the University of California, Berkeley in 2020. He then worked as an educator at Weights & Biases before joining @Full Stack Deep Learning, an online community and MOOC for building with ML. Sahar Mor Sahar is a Product Lead at @stripe with 15y of experience in product and engineering roles. At Stripe, he leads the adoption of LLMs and the Enhanced Issuer Network - a set of data partnerships with top banks to reduce payment fraud. Prior to Stripe he founded a document intelligence API company, was a founding PM in a couple of AI startups, including an accounting automation startup (Zeitgold, acq'd by Deel), and served in the elite intelligence unit 8200 in engineering roles. Sahar authors a weekly AI newsletter (AI Tidbits) and maintains a few open-source AI-related libraries (https://github.com/saharmor). Sarah Guo Sarah Guo is the Founder and Managing Partner at @Conviction, a venture capital firm founded in 2022 to invest in intelligent software, or "Software 3.0." Prior, she spent a decade as a General Partner at Greylock Partners. She has been an early investor or advisor to 40+ companies in software, fintech, security, infrastructure, fundamental research, and AI-native applications. Sarah is from Wisconsin, has four degrees from the University of Pennsylvania, and lives in the Bay Area with her husband and two daughters. She co-hosts the AI podcast "No Priors" with Elad Gil. Shyamala Prayaga Shyamala is a seasoned conversational AI expert. Having led initiatives across connected home, automotive, wearables - just to name a few, she's put her work on research into usability, accessibility, speech recognition, multimodal voice user interfaces, and has even been published internationally across publications like Forbes. Outside of her research, she's spent the last 18 years designing mobile, web, desktop, and smart TV interfaces and has most recently joined @NVIDIA to work on deep learning product suites. Willem Pienaar Willem is the creator of @Feast, the open-source feature store and a builder in the generative AI space. Previously Willem was an engineering manager at Tecton where he led teams in both their open source and enterprise initiatives. Before that Willem built the core ML systems and created the ML platform team at Gojek, the Indonesian decacorn. // Sign up for our Newsletter to never miss an event: https://mlops.community/join/ // Watch all the conference videos here: https://home.mlops.community/home/collections // Check out the MLOps Community podcast: https://open.spotify.com/show/7wZygk3mUUqBaRbBGB1lgh?si=242d3b9675654a69 // Read our blog: mlops.community/blog // Join an in-person local meetup near you: https://mlops.community/meetups/ // MLOps Swag/Merch: https://mlops-community.myshopify.com/ // Follow us on Twitter: https://twitter.com/mlopscommunity //Follow us on Linkedin: https://www.linkedin.com/company/mlopscommunity/
MLOps podcast #192 with Chris Van Pelt, CISO and co-founder of Weights & Biases, Enterprises Using MLOps, the Changing LLM Landscape, MLOps Pipelines sponsored by @WeightsBiases . // Abstract Chris, provides insights into his machine learning (ML) journey, emphasizing the significance of ML evaluation processes and the evolving landscape of MLOps. The conversation covers effective evaluation metrics, demo-driven development nuances, and the complexities of ML Ops pipelines. Chris reflects on his experience with Crowdflower, detailing its transition to Weights and Biases and stressing the early integration of security measures. The discussion extends to the transformative impact of ML on the tech industry, challenges in detecting subtle bugs, and the potential of open-source models and multimodal capabilities. // Bio Chris Van Pelt is a co-founder of Weights & Biases, a developer MLOps platform. In 2009, Chris founded Figure Eight/CrowdFlower. Over the past 12 years, Chris has dedicated his career optimizing ML workflows and teaching ML practitioners, making machine learning more accessible to all. Chris has worked as a studio artist, computer scientist, and web engineer. He studied both art and computer science at Hope College. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://wandb.ai/site --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Chris on LinkedIn: https://www.linkedin.com/in/chrisvanpelt/ Timestamps: [00:00] Chris' preferred coffee [00:33] Takeaways[03:50] Huge shout out to Weights & Biases for sponsoring this episode! [04:15] Please like, share, and subscribe to our MLOps channels! [04:25] CrowdFlower [07:02] Difference of CrowdFlower and Trajectory [09:13] Transition from CrowdFlower to Weights & Biases [13:05] Excel spreadsheets being passed around via email [15:45] Evolution of Weights & Biases [19:24] CISO role [22:23] Advise for easy wins [25:32] Transition into LLMs [27:36] Prompt injection risks on data [29:42] LLMs for New Personas [34:42] Iterative Value Evaluation Process [36:36] Iterating on New Release [39:31] Evaluation survey [43:21] Landscape of LLMs and its evolution [45:40] Conan O'Brien [46:48] Wrap up
MLOps podcast #191 with Gregory Kamradt, Founder of @DataIndependent, Building Defensible AI Apps sponsored by @MilvusVectorDatabase . // Abstract Demetrios engages in a captivating conversation with Gregory Kamradt, an AI visionary deeply immersed in technology and product development. The discussion spans various challenges businesses encounter in implementing AI, the transformative potential of AI in revolutionizing business processes, and the growth and possibilities associated with OpenAI. Gregory shares insights into his latest project, a smart companion app designed to analyze and summarize startup pitches. The episode unfolds as a rich source of knowledge, exploring diverse topics such as AI experimentation, the concept of an AI gateway, the future of finely tuned models for niche applications, and insights into the intricate landscape of AI within big tech, including Google's strategic direction and OpenAI's copyright protection measures. // Bio Greg has mentored thousands of developers and founders, empowering them to build AI-centric applications. By crafting tutorial-based content, Greg aims to guide everyone from seasoned builders to ambitious indie hackers. Greg partners with companies during their product launches, feature enhancements, and funding rounds. His objective is to cultivate not just awareness, but also a practical understanding of how to optimally utilize a company's tools. He previously led Growth @ Salesforce for Sales & Service Clouds in addition to being early on at Digits, a FinTech Series-C company. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://gregkamradt.com/ Greg Kamradt (Data Indy): https://www.youtube.com/@DataIndependent Milvus Vector Database: https://zilliz.com/what-is-milvus --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Greg on LinkedIn: https://www.linkedin.com/in/gregkamradt/ Timestamps: [00:00] Greg's preferred coffee [00:12] Takeaways [02:56] Quick word from our sponsor [04:22] DevDay [06:19] YouTube's unique perspective on the technological revolution [09:34] GPT assistance [13:36] AI Streamlining Fax Orders [18:13] AI Marketplace Dynamics: GPT vs. Specialized [22:04] Data Tooling Platform Challenges [27:17] The Shield against copyright [29:27] Llama Index vs OpenAI [31:56] DS Pie and Compiler Tangent [34:31] Orchestration Layer is dead! [36:49] Personalized AI Models: Understanding Integration [38:00] AI Defensibility [43:00] Green Field AI Opportunities [46:57] LLMs for live event pitch [53:38] Exciting content creation process [58:03] New context window benchmark [1:02:23] AI Gateway [1:04:35] Wrap up
MLOps podcast #190 with Ads Dawson, Senior Security Engineer at Cohere, Guarding LLM and NLP APIs: A Trailblazing Odyssey for Enhanced Security. // Abstract Ads Dawson, a seasoned security engineer at Cohere, explores the challenges and solutions in securing large language models (LLMs) and natural language programming APIs. Drawing on his extensive experience, Ads discusses approaches to threat modeling LLM applications, preventing data breaches, defending against attacks, and bolstering the security of these critical technologies. The presentation also delves into the success of the "OWASP Top 10 for Large Language Model Applications" project, co-founded by Ads, which identifies key vulnerabilities in the industry. Notably, Ads owns three of the top 10 vulnerabilities, including Training Data Poisoning, Sensitive Information Disclosure, and Model Theft. This OWASP Top 10 serves as a foundational resource for stakeholders in AI, offering guidance on using, developing, and securing LLM applications. Additionally, the session covers insider news from the AI Village's 'Hack the Future' | LLM Red Teaming event at Defcon31, providing insights into the inaugural Generative AI Red Teaming showdown and its significance in addressing security and privacy concerns amid the widespread adoption of AI. // Bio A mainly self-taught, driven, and motivated proficient application, network infrastructure & cyber security professional holding over eleven years experience from start-up to large-size enterprises leading the incident response process and specializing in extensive LLM/AI Security, Web Application Security and DevSecOps protecting REST API endpoints, large-scale microservice architectures in hybrid cloud environments, application source code as well as EDR, threat hunting, reverse engineering, and forensics. Ads have a passion for all things blue and red teams, be that offensive & API security, automation of detection & remediation (SOAR), or deep packet inspection for example. Ads is also a networking veteran and love a good PCAP to delve into. One of my favorite things at Defcon is hunting for PWNs at the "Wall of Sheep" village and inspecting malicious payloads and binaries. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://github.com/GangGreenTemperTatum OWASP Top 10 for Large Language Model Applications Core Team Member and Founder - https://owasp.org/www-project-top-10-for-large-language-model-applications/CoreTeam Fork for OWASP Top 10 for Large Language Model Applications - https://github.com/GangGreenTemperTatum/www-project-top-10-for-large-language-model-applications Security project: llmtop10.com --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ads on LinkedIn: https://www.linkedin.com/in/adamdawson0/ Timestamps: [00:00] Ads' preferred coffee [00:46] Takeaways [02:52] Please like, share, and subscribe to our MLOps channels! [03:11] Security and vulnerabilities [05:24] Work at Cohere and OWASP [08:11] Previous work vs LLMs Companies [09:46] LLM vulnerabilities [10:38] Good qualities to combat prompt injection problems [13:26] Data lineage [16:03] Red teaming [19:39] Freakiest LLM vulnerabilities [22:17] Severe Autonomy Concerns [25:13] Hallucinations [27:59] Prompt injection [29:15] Vector attacks to be recognized [32:02] LLMs being customed [33:18] Security changes due to maturity [38:17] OWASP Top 10 for Large Language Model Applications [44:31] Gandalf game [46:06] Prompt injection attack [49:46] Overlapping security [53:26] Data poisoning [56:57] Toxic data for LLMs [58:50] Wrap up
MLOps podcast #189 with Rohit Agarwal, CEO of Portkey.ai, Designing for Forward Compatibility in Gen AI. // Abstract For two whole years of working with a large LLM deployment, I always felt uncomfortable. How is my system performing? Are my users liking the outputs? Who needs help? Probabilistic systems can make this really hard to understand. In this talk, we'll discuss practical & implementable items to secure your LLM system and gain confidence while deploying to production. // Bio Rohit is the Co-founder and CEO of portkey.ai which is an FMOps stack for monitoring, model management, compliance, and more. Previously, he headed Product & AI at Pepper Content which has served ~900M generations on LLMs in production. Having seen large LLM deployments in production, he's always happy to help companies build their infra stacks on FM APIs or Open-source models. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://portkey.ai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Rohit on LinkedIn: https://www.linkedin.com/in/1rohitagarwal/ Timestamps: [00:00] Rohit's preferred coffee [00:15] Takeaways [03:22] Please like, share, and subscribe to our MLOps channels! [05:16] Rohit's current work [06:37] The Portkey landscape [09:13] Compute unit is no longer a Cloud resource, it's a Foundational Model [11:09] Hang-ups at high-scale models and how to combat them [15:22] Complexity of the Apps evolving [19:54] Rohit's working relationships with the agents [22:52] Fine-tuning reliability [24:38] Small language models can outperform larger ones [26:38] Market map at Portkey [34:37] AI Gateway [37:59] Worker Bee and Queen Bee [39:27] Security and Compliance [43:11] Idea of Data Mesh [45:57] Forward compatibility [49:59] Decoupling AI Gateway from the code [56:05] Hardest design decisions to make since creating Portkey [58:52] Wrap up
MLOps podcast #188 with Anand Das, Co-founder and CTO of Bito, Impact of LLMs on the Tech Stack and Product Development. // Abstract Anand and his team have developed a fascinating Chrome extension called "explain code" that has garnered significant attention in the tech community. They have expanded their extension to other platforms like Visual Studio code and Chat Brains, creating a personal assistant for code generation, explanation, and test case writing. // Bio Anand Das is the co-founder and CTO of Bito. Previously, he served as the CTO at Eyeota, which was acquired by Dun & Bradstreet for $165M in 2021. Anand also co-founded and served as the CTO of PubMatic in 2006, a company that went public on NASDAQ in 2020 (NASDAQ: PUBM). Anand has also held various engineering roles at Panta Systems, a high-performance computing startup led by the CTO of Veritas, as well as at Veritas and Symantec, where he worked on a variety of storage and backup products. Anand holds seven patents in systems software, storage software, advertising, and application software. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://bito.ai/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Anand on LinkedIn: https://www.linkedin.com/in/ananddas/ Timestamps: [00:00] Anand's preferred coffee [00:15] Takeaways [02:49] Please like, share, and subscribe to our MLOps channels! [03:08] Anand's tech background [10:06] Fun at Optimization Level [12:59] Trying all APIs [17:55] Models evaluation decision tree [22:51] Weights and Biases Ad [25:04] AI Stack that understands the code [28:27] Tools for the Guard Rails [33:23] Seeking solutions before presenting to LLM [38:46] Prompt-Driven Development Insights [40:16] Prompting best practices [42:51] Unneeded complexities [45:45] Cost-benefit analysis of buying GPUs [49:13] ML Build vs Buy [51:26] Best practices for debugging code assistant [54:58] Wrap up
MLOps podcast #187 with Faizaan Charania, Product Manager, AI at LinkedIn, Building Effective Products with GenAI. // Abstract Faizaan outlines his AI product development approach, starting broadly and refining details with tech leads, emphasizing the value of a simplified MVP. He also explores integrating generative AI, highlighting its role in enhancing user experiences through LLMs. In this discussion, Faizaan shares wisdom on feedback integration, user trust, and the collaboration challenges between product managers and AI teams. Let's delve into evaluating AI-driven experiences and the complexities that arise in this dynamic landscape! // Bio Faizaan is an AI Product lead at LinkedIn working on Personalization and Generative AI use cases for Creators and Conversations on LinkedIn. He's been in the field of machine learning for 8+ years now he started as a research assistant eventually transitioning to being a Product Manager during his time at Yahoo. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Faizaan on LinkedIn: https://www.linkedin.com/in/faizaan-charania/ Timestamps: [00:00] Faizaan's preferred beverage [00:22] Takeaways [02:54] The Bollywood actor [05:23] Faizaan's in tech [07:45] Technical pieces to learn about before working at LinkedIn [09:23] Tech Team Data Strategy [12:01] Gradual vs. Advanced ML Implementation [13:36] Shipping on time [14:11] Thoughts on building products with AI [18:20] Push and pull mechanism [21:47] Costs and Choices with AI Models [25:06] AI ROI Evaluation [27:02] Thoughts on open source [28:17] Building Generative AI focus [31:50] Prompts and Anomalies [34:57] Where to have a human in the loop [35:45] Problem-driven AI Tool [37:56] Creator of AI-generated post on LinkedIn [39:50] Product Impact on AI Democratization [41:15] Distinct signals to measure ROI [44:38] PMs learning AI while ML teams learn product [47:22] Gotchas seen when adding a new AI feature [50:00] Evaluation Challenges in Responses [51:55] Who's more confident? [52:55] Wrap up
MLOps podcast #186 with Mike Del Balso, CEO & Co-founder of Tecton and Josh Wills, Angel Investor, The Future of Feature Stores and Platforms. // Abstract Mike and Josh discuss creating templates and working at a detailed level, exploring Tecton's potential for sharing fraud and third-party features. They focus on technical aspects like data handling and optimizing models, emphasizing the significance of quality data for AI systems and the necessity for cohesive feature infrastructure in reaching production stages. // Bio Mike Del Balso Mike is the co-founder of Tecton, where he is focused on building next-generation data infrastructure for Operational ML. Before Tecton, Mike was the PM lead for the Uber Michelangelo ML platform. He was also a product manager at Google where he managed the core ML systems that power Google’s Search Ads business. Josh Wills Josh Wills is an angel investor specializing in data and machine learning infrastructure. He was formerly the head of data engineering at Slack, the director of data science at Cloudera, and a software engineer at Google. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Mike on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/Connect with Josh on LinkedIn: https://www.linkedin.com/in/josh-wills-13882b/ Timestamps: [00:00] Introduction to Mike [01:45] Takeaways [03:32] Features of the new paradigm of ML and LLMs [06:00] D. Sculley's papers [13:05] The birth of Feature Store [17:06] Data Pipeline Challenges Addressed [20:00] Operationalizing [26:50] Feature Store Challenges [30:26] Z access [36:23] Addressing Technical Debt Challenges [37:27] Real-Time vs. Batch Processing [47:10] Feature Store Evolution: Apache Iceberg [49:59] Feature Platform: Dedicated Query Engine [54:04] The bottleneck [56:00] LLMs, Feature Stores Overview [1:00:20] Vector databases [1:06:15] Workflow Templating Efficiency [1:08:35] Gamification suggestion for Tecton [1:10:25] Wrap up
MLOps podcast #185 with Luigi Patruno, VP of Data Science at 2U, Inc, Lessons on Data Science Leadership. // AbstractPicture this: you've got data products to manage, and you're in charge of a team. It's not all sunshine and rainbows, right? Luigi dives into the nitty-gritty of the challenges - from juggling data projects to wrangling the team dynamics. It's a real adventure, let me tell you! // Bio Luigi Patruno is a results-driven data science leader passionate about identifying value-add business opportunities and converting these into analytical solutions that deliver measurable business outcomes. As a leader he focuses on defining strategic vision and, through motivation and discipline, driving teams of highly quantitative data scientists, machine learning engineers, and product managers to achieve extraordinary results. He is currently the VP of Data Science at 2U, where he leads the data science department focused on optimizing business operations through advanced analytics, experimentation, and machine learning. He enjoys teaching others how to leverage data science to improve their businesses through public speaking, teaching courses, and writing online at MLinProduction.com. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://mlinproduction.com/ YouTube channel: https://www.youtube.com/playlist?list=PLBLnN4jzkyqkjLIRpDNZcsG7TMMEk9Asa High Output Management book by Andrew Grove: https://www.amazon.nl/-/en/Andrew-S-Grove/dp/0679762884The One Minute Manager by Kenneth Blanchard Ph.D. and Spencer Johnson M.D.: https://www.amazon.com/Minute-Manager-Kenneth-Blanchard-Ph-D/dp/074350917X --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Luigi on LinkedIn: https://www.linkedin.com/in/luigipatruno91/ Timestamps: [00:00] Luigi's preferred coffee [00:30] Takeaways [03:04] Being practical [05:44] Data-Driven Decision-Making in Management [12:53] Recent Team Win [14:43] The perfect storm [20:22] Change Management and ROI [25:09] Change Management: Navigating Resistance [29:59] Clarifying North Star Communication [36:24] OKRs in Data Science [40:47] Success Likelihood in Business [45:08] Bus problem solution [49:25] Data Science-Platform Collaboration [53:19] Decentralized Platforms Explained [54:38] Data Platform Architecture Overview [57:14] Incentives for Team Motivation [1:09:45] The blind spots [1:12:22] Wrap up
MLOps podcast #184 with Richa Sachdev, Executive Director- Data Operations and Automation at JP Morgan Chase, Data Platforms in MLOps: Translating Business Goals into Product Decisions.
// Abstract
Richa, with her background in software engineering and experience in the financial sector, shares her insights on optimizing the end-user experience and the importance of understanding business goals and metrics. She discusses her journey in converting legacy applications, working with data platforms, and the challenges of integrating different databases. Richa also explores the role of automation in streamlining processes and improving customer interactions in the reward space. Join us as we unravel the fascinating world of MLOps and uncover the strategies and technologies that drive success in this ever-evolving field.
// Bio
A passionate and impact-driven leader whose expertise spans leading teams, architecting ML and data-intensive applications, and driving enterprise data strategy.
Richa has worked for a Tier A Start-up developing feature platforms and in financial companies, leading ML Engineering teams to drive data-driven business decisions.
Richa enjoys reading technical blogs focussed on system design and plays an active role in the MLOps Community.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://www.youtube.com/watch?v=i0To3DeHGuU
https://www.youtube.com/watch?v=tAOf2lVQUY4
https://www.youtube.com/watch?v=cXanVyaannQ
https://www.youtube.com/watch?v=2aWSsL24fv8
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Richa on LinkedIn: https://www.linkedin.com/in/richasachdev/
Timestamps:
[00:00] Richa's preferred coffee
[02:09] Takeaways
[04:26] Richa's background to data
[08:55] Prescriptive, Descriptive, and Predictive Data
[11:50] Data Engineering Perspectives & Setup
[17:34] Structured and Unstructured data
[21:01] Richa's day-to-day at Chase
[23:52] Figure out the business needs before the cool tech
[26:46] Importance of business metrics
[30:43] Optimizing end-user experience and trade-offs
[36:06] Exhausting creativity in finding solutions
[37:40] Consider faster implementation and increased ROI
[40:20] Banks still using COBOL
[41:17] Learning and growing as a versatile leader
[42:04] Wrap up
MLOps podcast #183 with Ketan Umare, CEO of Union.AI, MLOps vs ML Orchestration co-hosted by Stephen Batifol. // Abstract Let's explore the relationship between Union and Flyte, emphasizing the significance of community-driven development and the challenge of balancing feature requests with security considerations. This conversation highlights the importance of real-time data and secure data handling in orchestrating machine learning models. The Flyte community's empathy and support for newcomers underscore the community's value in democratizing machine learning, making it more accessible and efficient for a broader audience. // Bio Ketan Umare is the CEO and co-founder at Union.ai. Previously he had multiple Senior roles at Lyft, Oracle, and Amazon ranging from Cloud, Distributed storage, Mapping (map-making), and machine-learning systems. He is passionate about building software that makes engineers' lives easier and provides simplified access to large-scale systems. Besides software, he is a proud father, and husband, and enjoys traveling and outdoor activities. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://union.ai/ Flyte: https://flyte.org/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stephen-batifol/ Connect with Ketan on LinkedIn: https://www.linkedin.com/in/ketanumare/ Timestamps: [00:00] Ketan's preferred coffee [01:05] Takeaways [03:08] Please like, share, and subscribe to our MLOps channels! [03:15] Shout out to Ketan and UnionAI for sponsoring this episode! [04:23] Orchestration recent changes [07:51] Community with Flyte [11:26] ML orchestration [15:40] 50/50 is generous [20:06] Real-time ML [21:15] Over engineering without benefits [23:20] Balancing everything [27:40] Union verse Flyte [32:52] High value features of Union AI at the back of Flyte [40:18] Building LLM infrastructure [45:30] Traditional ML is the whole prompting [46:46] LLMs to evaluating prompts [48:55] Wrap up
MLOps podcast #182 with GetYourGuide's Jean Machado, DataScience Manager, Meghana Satish, MLOps Engineer, Olivia Houghton, Machine Learning Operations Engineer, Theodore Meynard, Data Science Manager, MLOps@GetYourGuide. // Abstract Join a team to talk about the journey of GYG with MLOps. From the conception of their platform to the creation of the MLOps engineer role and to their current stack state. // Bio Jean Machado Jean Carlo Machado is a DataScience Manager at GetYourGuide for the Growth Data Products team and the Machine Learning Platform Team. He is privileged to be able to work on turning ideas in data scinece from inception to production. Before GYG Jean was working in a startup in Brazil building its infrastructure from the ground up. Jean also likes community building and using technology for social good. Meghana Satish Meghana Satish is currently working as an MLOps Engineer at GetYourGuide. She has previously held positions at Amazon AWS in Berlin and Microsoft IT in Hyderabad. In addition to her career in technology, Meghana is also a talented singer, dancer, and yoga practitioner. Olivia Houghton Olivia has been working as an MLOps engineer at GetYourGuide for the past year and a half or so. Olivia's main work is in building and managing their activity ranking service. Theodore Meynard Theodore Meynard, Data Science Manager at GetYourGuide, leads the evolution of their ranking algorithm, enriching customer experiences. His hands-on journey from data scientist to leader has honed his expertise in MLOps and real-time ML. Beyond work, he's a co-organizer of PyData Berlin, underlining his commitment to community and collaborative learning. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://investing1012dot0.substack.com/ The Openness of AI report: https://research.contrary.com/reports/the-openness-of-ai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jean on LinkedIn: https://www.linkedin.com/in/jean-carlo-machado-53b15977/ Connect with Meghana on LinkedIn: https://www.linkedin.com/in/meghana-satish-2a825282/?originalSubdomain=de Connect with Olivia on LinkedIn: https://www.linkedin.com/in/oliviaphoughton/ Connect with Theodore on LinkedIn: https://www.linkedin.com/in/theodore-meynard/
Timestamps: [00:00] GetYourGuide team's preferred coffee [00:55] Takeaways [02:20] Shout out to Berlin MLOps Community [02:38] Please like, share, and subscribe to our MLOps channels! [03:39] The GetYourGuide platform [05:45] GetYourGuide use cases [11:51] Strong Leadership Vision [13:59] Creating rituals [16:55] Feedback on the loop for improvements [18:35] Different components of GetYourGuide's ML Platform [21:04] V2 service templates [24:26] Biggest pain points [27:02] Feature flags [30:51] Data foundation [36:25] Data Testing [39:53] Cross-team Tool Adoption Process [44:59] Regrets about design decisions made in the past [47:53] What's next for the platform with LLMs? [52:49] Non-data scientists suggesting use cases, language flexibility [55:14] DevSecOps team's AI study group ideation [59:25] Experiments in growth data products, marketing split [1:01:47] Shout out to the Berlin MLOps Community! [1:03:31] Wrap up
MLOps podcast #181 with Kyle Harrison, General Partner at Contrary, The Centralization of Power in AI. // Abstract Kyle Harrison delves into the limitations imposed by language, underscoring how it can impede our grasp and manipulation of reality while stressing the critical need for improved language model performance for real-time applications. He further explores the perils of centralizing power in AI, with a specific focus on the "Openness of AI", where concerns about privacy are brought to the forefront, prompting his call for businesses to reconsider their reliance on it. The discussion also traverses the evolving landscape of AI, drawing comparisons between prominent machine learning frameworks such as TensorFlow and PyTorch. Notably, the episode underscores the vital role of open-source initiatives within the AI community and highlights the unexpected involvement of Meta in driving open-source development. // Bio Kyle Harrison is a General Partner at Contrary, where he leads Series A and growth-stage investing. He joined Contrary from Index where he was a Partner, and before that he was a growth investor at Coatue. His portfolio includes iconic startups and public companies including Ramp, Replit, Cohere, Snowflake, and Databricks. He also regularly shares his analysis on the venture capital landscape via his Substack Investing 101. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://investing1012dot0.substack.com/ The Openness of AI report: https://research.contrary.com/reports/the-openness-of-ai --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Kyle on LinkedIn: https://www.linkedin.com/in/kyle-harrison-9274b278/ Timestamps: [00:00] Kyle's preferred beverage [00:20] Takeaways [03:52] Hype in technology space [09:20] Application Layer Revenue [14:44] Stability AI Lawsuit [18:08] Concern over concentration of power in AI [20:20] Transparency concerns [23:35] Open Source AI [25:57] To use or not to use Open AI [30:51] Lack of technical expertise and business-building capabilities [35:09] AI Transparency and Accountability [37:50] Traditional ML [41:47] Finding a unique approach [45:41] AGI limitations [47:43] Using Agents [49:46] Agents getting past demos [54:39] Tech Challenges & Hoverboard Dreams [58:04] Both AI hype and skepticism are foolish [01:27] Wrap up
MLOps podcast #180 with Sachin Abeywardana Deep Learning Engineer at Canva AI, Adventures in Building CLIP and Other (Largeish) Language Models sponsored by Prem AI. // Abstract Sachin takes us on an adventure, sharing insights on the pitfalls of not understanding the broader product and the importance of incorporating AI and machine learning capabilities. From the use of AI models to grammar correction and code generation to the fascinating Clip model and the challenges of balancing work and family life, this episode promises to be both informative and thought-provoking. // Bio Sachin is the father of two beautiful children. He completed his PhD in Bayesian Machine Learning at University of Sydney in 2015. In 2016 he discovered Deep Learning and hasn't looked back. He currently works as a Senior Machine Learning Engineer at Canva and is mainly focusing on NLP problems. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Sachin Blogs: https://sachinruk.github.io/blog.html https://sachinruk.github.io/blog/ Graph ML link: http://web.stanford.edu/class/cs224w/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sachin on LinkedIn: https://www.linkedin.com/in/sachinabeywardana/ Timestamps: [00:00] Sachin's preferred beverage [00:26] Takeaways [02:30] Chat GPT user [05:58] Understanding on reliable Agents [08:10] Sachin's background [12:45] Staying at Deep Learning [16:17] Recommendation or Lead Scoring [17:36] Vector database [19:00] Sachin's blogs [23:26] The cap people [26:10] Pursuing business case [27:33] Canva [31:16] Incorporating AI and Machine Learning [32:17] Sponsor Ad [38:22] Eliminating unnecessary steps [39:00] Interacting with the product team [43:04] Criticisms on the current architecture limitations [45:58] Insufficient exploration of Transformers [47:42] Explaining GraphML [52:35] Fine-tuning ChatGPT2 [57:54] Leading ML Engineers and teams [59:40] Being practical with Math [1:05:52] Wrap up
MLOps Coffee Sessions #179 with Shahul Es, All About Evaluating LLM Applications. // Abstract Shahul Es, renowned for his expertise in the evaluation space and the creator of the Ragas Project. Shahul dives deep into the world of evaluation in open source models, sharing insights on debugging, troubleshooting, and the challenges faced when it comes to benchmarks. From the importance of custom data distributions to the role of fine-tuning in enhancing model performance, this episode is packed with valuable information for anyone interested in language models and AI. // Bio Shahul is a data science professional with 6+ years of expertise and has worked in data domains from structured, NLP to Audio processing. He is also a Kaggle GrandMaster and code owner/ ML of the Open-Assistant initiative that released some of the best open-source alternatives to ChatGPT. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links All about evaluating Large language models blog: https://explodinggradients.com/all-about-evaluating-large-language-models Ragas: https://github.com/explodinggradients/ragas --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Shahul on LinkedIn: https://www.linkedin.com/in/shahules/ Timestamps: [00:00] Shahul's preferred coffee [00:20] Takeaways [01:46] Please like, share, and subscribe to our MLOps channels! [02:07] Shahul's definition of Evaluation [03:27] Evaluation metrics and Benchmarks [05:46] Gamed leaderboards [10:13] Best at summarizing long text open-source models [11:12] Benchmarks [14:20] Recommending evaluation process [17:43] LLMs for other LLMs [20:40] Debugging failed evaluation models [24:25] Prompt injection [27:32] Alignment [32:45] Open Assist [35:51] Garbage in, garbage out [37:00] Ragas [42:52] Valuable use case besides Open AI [45:11] Fine-tuning LLMs [49:07] Connect with Shahul if you need help with Ragas @Shahules786 on Twitter [49:58] Wrap up
MLOps Coffee Sessions #178 with Stephen Batifol, Building an ML Platform: Insights, Community, and Advocacy. // Abstract Discover how Wolt onboard data scientists onto the platform and build a thriving internal community of users. Stephen's firsthand experiences shed light on the importance of developer relations and how they contribute to making data scientists' lives easier. From top-notch documentation to getting-started guides and tutorials, the internal platform at Wolt prioritizes the needs of its users. // Bio From Android developer to Data Scientist to Machine Learning Engineer, Stephen has a wealth of software engineering experience at Wolt. He believes that machine learning has a lot to learn from software engineering best practices and spends his time making ML deployments simple for other engineers. Stephen is also a founding member and organizer of the MLOps.community Meetups in Berlin. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stephen-batifol/ Timestamps: [00:00] Stephen's preferred coffee [00:32] Takeaways [01:35] Please like, share, and subscribe to our MLOps channels! [03:00] Creating his own team! [04:44] DevRel [06:32] The door dash of Europe [11:28] Data platform underneath [12:55] Cellular core deployment uses open source [14:21] Alibi [16:08] Kafka [16:59] Selling points to data scientists [20:05] Language models concerns of data scientists [22:12] Incorporating LLMs into the business [23:55] Feedback from data scientists and end users [27:37] User surveys [30:11] Evangelizing and giving talks [35:25] Tech Hub Culture in Berlin [38:38] Kubernetes lifestyle [42:55] Interacting with SREs [45:28] Wrap up
MLOps Coffee Sessions #176 with Vin Vashishta, Collaboration and Strategy. // Abstract From the significance of technical strategists to the crucial role of product managers with a deep understanding of data and AI products, Vin shares invaluable insights on fostering collaboration, driving strategy, and maximizing the potential of data within organizations. Join us as we explore the importance of becoming multipliers in our fields, the power of effective strategy in leveraging data, and the opportunities that lie in the generative AI era. // Bio Vin's background is in applied data science. He is the founder of V Squared, one of the oldest and smallest data science consulting companies in the world. They help businesses monetize data and AI. Vin is the author of From Data To Profit. He teaches technical strategy and data and AI product management. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.datascience.vin/ From Data To Profit: How Businesses Leverage Data to Grow Their Top and Bottom Lines book: https://www.amazon.com/Data-Profit-Businesses-Leverage-Bottom/dp/1394196210/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vin on LinkedIn: https://www.linkedin.com/in/vineetvashishta/ Timestamps: [00:00] Vin's preferred coffee [00:14] Takeaways [02:02] Please like, share, and subscribe to our MLOps channels! [02:28] Recent ideas of Vin [05:09] Understanding the business value of any project [08:30] Generative AI making things faster [14:29] Strategy in practice [20:19] Practicality and Credibility of Strategists [22:42] Coming soon!!! LLMs in Production Conference Panel Part III [27:48] Becoming a Multiplier [29:03] The AI Product Manager [35:12] Successful monetization and integration of technologies [37:48] Justifying the ROI of LLMs [44:59] Adding that extra value [49:52] Read Vin's book linked above! [50:35] Wrap up
Sign up for our next LLM in production conference: https://go.mlops.community/prodiii
#180 with LLMs in Production Conference part 2 Ux of a LLM User Panel, Misty Free, Dina Yerlan, and Artem Harutyunyan hosted by Innovation Endeavors' Davis Treybig. // Abstract Explore different approaches to interface design, emphasizing the significance of crafting effective prompts and addressing accuracy and hallucination issues. Discover some strategies for improving latency and performance, including monitoring, scaling, and exploring emerging technologies. // Bio Misty Free Misty Free is a product manager at Jasper, where she focuses on supercharging marketers with speed and consistency in their marketing campaigns, with the power of AI. Misty has also collaborated with Stability and OpenAI to offer AI image generation within Jasper. She approaches product development with a "jobs-to-be-done" mindset, always starting with the "why" behind any need, ensuring that customer pain points are directly addressed with the features shipped at Jasper. In her free time, Misty enjoys crocheting amigurumi, practicing Spanish on Duolingo, and spending quality time with her family. Misty will be on a panel sharing her insights and experiences on the real-world use cases of LLMs. Davis Treybig Davis is a partner at Innovation Endeavors, an early-stage venture firm focused on teams solving hard technical & engineering problems. He personally focuses on computing infrastructure, AI/ML, and data. Dina Yerlan Head of Product, Generative AI Data at Adobe Firefly (family of foundation models for creatives). Artem Harutyunyan Artem is the Co-Founder & CTO at Bardeen AI. Prior to Bardeen, he was in engineering and product roles at Mesosphere and Qualys, and before that, he worked at CERN. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.angellist.com/venture/relay Foundation by Isaac Asimov: https://www.amazon.com/Foundation-Isaac-Asimov/dp/0553293354 AngelList Relay blog: https://www.angellist.com/blog/introducing-angellist-relay --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Davis on LinkedIn: https://www.linkedin.com/in/davistreybig/ Connect with Misty on LinkedIn: https://www.linkedin.com/in/misty-miglorin/ Connect with Dina on LinkedIn: https://www.linkedin.com/in/dinayerlan/ Connect with Artem on LinkedIn: https://www.linkedin.com/in/artemharutyunyan/
MLOps Coffee Sessions #175 with Lamia Youseff, From Virtualization to AI Integration. // Abstract Lamia discusses how both Fortune 500 companies and SMBs lack the knowledge and capabilities to identify which use cases in their systems can benefit from AI integration. She emphasizes the importance of helping these companies integrate AI effectively and acquire the necessary capabilities to stay competitive in the market. // Bio By way of an introduction, Dr. Lamia Youseff has been working in AI / ML for ~25 years, first in academia (MIT, Stanford, UCSB), then large tech (Google, Microsoft, Apple, and Facebook), and most recently with startups in Generative AI. She is currently the executive director of JazzComputing, a Visiting Research Scientist at Stanford University in Computer Science and AI, and a research affiliate with MIT Computer Science and Artificial Intelligence Lab (CSAIL). Dr. Youseff earned her Ph.D. in computer science by studying computationally intensive workloads (such as AI / ML and HPC / Scientific Codes) and has built/led several AI teams as an executive and leader at large tech companies over the years (Google, Facebook, Microsoft, and Apple). She also earned her Master's in business management, strategy, and leadership from Stanford Graduate School of Business (GSB), where she is a guest lecturer today. Dr. Youseff regularly writes and speaks about AI and Machine Learning evolution at CIO/CTO/CEO summits. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Lamia on LinkedIn: https://www.linkedin.com/in/lyouseff/ Timestamps: [00:00] Lamia's preferred coffee [01:12] Takeaways [03:00] Please like, share, and subscribe to our MLOps channels! [03:20] Lamia's background [09:52] Getting into Google Cloud [13:10] The Google Cloud project [16:38] The world before Kubernetes [19:25] Evolution of virtualization [23:20] Cloud evolution [28:13] Kubernetes enables the ecosystem [32:38] Multiple systems for machine learning [34:40] Standardization to a greater good [39:50] Complexity and pain points of ML in production [46:26] JazzComputing [50:33] Bridging gaps in AI implementation and investment [51:19] Wrap up
MLOps Coffee Sessions #178 with LLMs in Production Conference part 2 LLM on K8s Panel, Manjot Pahwa, Rahul Parundekar, and Patrick Barker hosted by Outerbounds, Inc.'s Shrinand Javadekar. // Abstract Large Language Models require a new set of tools... or do they? K8s is a beast and we like it that way. How can we best leverage all the battle-hardened tech that K8s has to offer to make sure that our LLMs go brrrrrrr. Let's talk about it in this chat. // Bio Shrinand Javadekar Shri Javadekar is currently an engineer at Outerbounds, focussed on building a fully managed, large-scale platform for running data-intensive ML/AI workloads. Earlier, he spent time trying to start an MLOps company for which he was a co-founder and head of engineering. He led the design, development, and operations of Kubernetes-based infrastructure at Intuit, running thousands of applications, built by hundreds of teams and transacting billions of $$. He has been a founding engineer of the Argo open-source project and also spent precious time at multiple startups that were acquired by large organizations like EMC/Dell and VMWare. Manjot Pahwa Manjot is an investor at Lightspeed India and focuses on SaaS and enterprise tech. She has had an operating career of over a decade within the space of fintech, SaaS, and developer tools spanning various geos such as the US, Singapore, and India. Before joining Lightspeed, Manjot headed Stripe in India, successfully obtaining the payment aggregator license, growing the team from ~10 to 100+, and driving acquisitions in the region during that time. Rahul Parundekar Rahul has 13+ years of experience building AI solutions and leading teams. He is passionate about building Artificial Intelligence (A.I.) solutions for improving the Human Experience. He is currently the founder of A.I. Hero - a platform to help you fix and enrich your data with ML. At AI Hero, he has also been a big proponent of declarative MLOps - using Kubernetes to operationalize the training and serving lifecycle of ML models and has published several tutorials on his Medium blog. Before AI Hero, he was the Director of Data Science (ML Engineering) at Figure-Eight (acquired by Appen), a data annotation company, where he built out a data pipeline and ML model serving architecture serving 36 models (NLP, Computer Vision, Audio, etc.) and traffic of up to 1M predictions per day. Patrick Barker Patrick started his career in Big Data back when that was cool, then moved into Kubernetes near its inception. He has put major features into the Kubernetes API and built several platforms on top of it. In recent years he has moved into AI, with a focus on distributed machine learning. He is now working with a startup to reshape the world of AI agents. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.angellist.com/venture/relay Foundation by Isaac Asimov: https://www.amazon.com/Foundation-Isaac-Asimov/dp/0553293354 AngelList Relay blog: https://www.angellist.com/blog/introducing-angellist-relay --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Shri on LinkedIn: https://www.linkedin.com/in/shrijavadekar/ Connect with Manjot on LinkedIn: https://www.linkedin.com/in/manjotpahwa/ Connect with Rahul on LinkedIn: https://www.linkedin.com/in/rparundekar/ Connect with Patrick on LinkedIn: https://www.linkedin.com/in/patrickbarkerco/
MLOps Coffee Sessions #174 with Michelle Marie Conway, Harnessing MLOps in Finance: Bringing Statistical Models to Life for Positive Impact, co-hosted by Stephen Batifol. // Abstract Michelle Marie Conway joins hosts Stephen Batifol and Demetrios to share their insights and experiences in the tech industry. Michelle emphasizes the importance of constant learning and adaptation in the rapidly changing tech industry. They discuss the need to stay up to date with the latest documentation, understand code logic, and be mindful when writing code. Michelle also reflects on their experiences as one of the few women in their university math class and often being the only woman on their team in the workplace. They discuss the need for more girls to pursue STEM subjects in schools and the importance of allies in the workplace. Additionally, Michelle explores the benefits and challenges of AI tools, sharing their experiences with tools like Gen AI and Chat GPT. While AI tools enhance productivity, Michelle also acknowledges the limitations of these tools in more technical tasks and the continued reliance on developer resources. This episode offers valuable insights into the importance of continuous learning, gender diversity in STEM, and the potential of AI tools in the field of MLOps. // Bio As an Irish woman who relocated to London after completing her university studies in Dublin, Michelle spent the past 12 years carving out a career in the data and tech industry. With a keen eye for detail and a passion for innovation, She has consistently leveraged my expertise to drive growth and deliver results for the companies she worked for. As a dynamic and driven professional, Michelle is always looking for new challenges and opportunities to learn and grow, and she's excited to see what the future holds in this exciting and ever-evolving industry. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stephen-batifol/ Connect with Michelle on LinkedIn: https://www.linkedin.com/in/michelle-conway-40337432 Timestamps: [00:00] Michelle's preferred coffee [02:04] Takeaways [05:18] Please like, share, and subscribe to our MLOps channels! [06:18] Michelle's journey in tech [07:49] Engineering best practices [09:38] Getting comfortable with the hump [11:22] Clean coding fundamentals [13:29] Working with the people [14:09] GCP migration [18:00] GCP migration length of journey [18:38] Moving data focus [19:18] Effectiveness of running 2 systems [21:00] Dealing with discrepancies [22:15] Using Nexus [24:04] Migrating data from Teradata to BigQuery, strict security [28:48] Hiring new people [30:56] Securely managing financial data with millions of customers [32:30] When things go wrong [35:08] Finding the root cause [36:28] Dealing with the producers' problems [40:46] Rapid tech evolution constant learning [44:44] Teaching Python, using Gen AI for tasks [46:34] Dealing with LLMs use cases [49:15] Dealing with stakeholders and MLOps teams [51:17] Having a translator [52:18] Being a woman in the tech industry [55:11] Encourage more girls in STEM, support women [56:36] Women in the conversation on tech and female representation [1:03:49] Wrap up
MLOps Coffee Sessions #176 with MLOps vs. LLMOps Panel, Willem Pienaar, Chris Van Pelt, Aparna Dhinakaran, and Alex Ratner hosted by Richa Sachdev. // Abstract What do MLOps and LLMOps have in common? What has changed? Are these just new buzzwords or is there validity in calling this ops something new? // Bio Richa Sachdev A passionate and impact-driven leader whose expertise spans leading teams, architecting ML and data-intensive applications, and driving enterprise data strategy. Richa has worked for a Tier A Start-up developing feature platforms and in financial companies, leading ML Engineering teams to drive data-driven business decisions. Richa enjoys reading technical blogs focussed on system design and plays an active role in the MLOps Community. Willem Pienaar Willem is the creator of Feast, the open-source feature store and a builder in the generative AI space. Previously Willem was an engineering manager at Tecton where he led teams in both their open source and enterprise initiatives. Before that Willem built the core ML systems and created the ML platform team at Gojek, the Indonesian decacorn. Chris Van Pelt Chris Van Pelt is a co-founder of Weights & Biases, a developer MLOps platform. In 2009, Chris founded Figure Eight/CrowdFlower. Over the past 12 years, Chris has dedicated his career optimizing ML workflows and teaching ML practitioners, making machine learning more accessible to all. Chris has worked as a studio artist, computer scientist, and web engineer. He studied both art and computer science at Hope College. Aparna Dhinakaran Aparna Dhinakaran is the Co-Founder and Chief Product Officer at Arize AI, a pioneer and early leader in machine learning (ML) observability. A frequent speaker at top conferences and thought leader in the space, Dhinakaran was recently named to the Forbes 30 Under 30. Before Arize, Dhinakaran was an ML engineer and leader at Uber, Apple, and TubeMogul (acquired by Adobe). During her time at Uber, she built several core ML Infrastructure platforms, including Michelangelo. She has a bachelor’s from UC Berkeley's Electrical Engineering and Computer Science program, where she published research with Berkeley's AI Research group. She is on a leave of absence from the Computer Vision Ph.D. program at Cornell University. Alex Ratner Alex Ratner is the co-founder and CEO at Snorkel AI, and an Affiliate Assistant Professor of Computer Science at the University of Washington. Prior to Snorkel AI and UW, he completed his Ph.D. in CS advised by Christopher Ré at Stanford, where he started and led the Snorkel open source project, and where his research focused on defining and forwarding the concept of “data-centric AI”, the idea that labeling and developing data is the new center of the AI development workflow. His academic work focuses on data-centric AI and related topics in data management and statistical learning techniques, and applications to real-world problems in medicine, science, and more. Previously, he earned his A.B. in Physics from Harvard University. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Richa on LinkedIn: https://www.linkedin.com/in/richasachdev/ Connect with Willem on LinkedIn: https://www.linkedin.com/in/willempienaar/ Connect with Chris on LinkedIn: https://www.linkedin.com/in/chrisvanpelt/ Connect with Aparna on Twitter: https://www.linkedin.com/in/aparnadhinakaran/ Connect with Alex on Twitter: https://www.linkedin.com/in/alexander-ratner-038ba239/
MLOps Coffee Sessions #173 with Beyang Liu, Building Cody, an Open Source AI Coding Assistant. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract Root about the development of Cody, an open-source AI coding assistant. Cody empowers developers to query and comprehend code within codebases through the integration of robust language model capabilities. Sourcegraph tackles the intricacies of understanding intricate codebases by creating comprehensive code maps and employing AI for advanced search functionalities. Cody harnesses the potential of AI to offer features such as code exploration, natural language queries, and AI-powered code generation, augmenting developer productivity and code comprehension. // Bio Beyang Liu is the CTO and Co-founder of Sourcegraph. Prior to Sourcegraph, Beyang was an engineer at Palantir Technologies building large-scale data analysis tools for Fortune 500 companies with large, complex codebases. Beyang studied computer science at Stanford, where he discovered his love for compilers and published some machine learning research as a member of the Stanford AI Lab. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://beyang.com --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Beyang on LinkedIn: https://www.linkedin.com/in/beyang-liu/ Timestamps: [00:00] Beyang's preferred coffee [00:19] Takeaways [01:25] Please like, share, and subscribe to our MLOps channels! [01:48] Beyang background before Sourcegraph [03:10] War stories [04:30] Technological tool solution [06:41] Landscape change in the past 10 years [09:32] Code search engine evolution [16:28] Vector databases [17:40] Actual tech breakdown [19:52] Incorporating AI into products amid organizational challenges [25:39] Breaking down Cody [28:04] Context fetching [30:44] AI replicating human code understanding? [36:22] Key for software creation [40:26] Speak the language [42:20] Leveraging LLMs [44:18] Low code, no code movement [47:54] Reliability issues amongst agents [53:12] LLMs used in code and chat generation [56:12] Dealing with rate limits and followers or failovers [57:33] Unnecessary comparison [1:00:26] Wrap up
MLOps Coffee Sessions #174 with Evaluation Panel, Amrutha Gujjar, Josh Tobin, and Sohini Roy hosted by Abi Aryan. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract Language models are very complex thus introducing several challenges in interpretability. The large amounts of data required to train these black-box language models make it even harder to understand why a language model generates a particular output. In the past, transformer models were typically evaluated using perplexity, BLEU score, or human evaluation. However, LLMs amplify the problem even further due to their generative nature thus making them further susceptible to hallucinations and factual inaccuracies. Thus, evaluation becomes an important concern. // Bio Abi Aryan Machine Learning Engineer @ Independent Consultant Abi is a machine learning engineer and an independent consultant with over 7 years of experience in the industry using ML research and adapting it to solve real-world engineering challenges for businesses for a wide range of companies ranging from e-commerce, insurance, education and media & entertainment where she is responsible for machine learning infrastructure design and model development, integration and deployment at scale for data analysis, computer vision, audio-speech synthesis as well as natural language processing. She is also currently writing and working in autonomous agents and evaluation frameworks for large language models as a researcher at Bolkay. Amrutha Gujjar CEO & Co-Founder @ Structured Amrutha Gujjar is a senior software engineer and CEO and co-founder of Structured, based in New York. With a Bachelor of Science in Computer Science from the University of Washington's Allen School of CSE, she brings expertise in software development and leadership to my work.
Connect with Amrutha on LinkedIn to learn more about her experience and discuss exciting opportunities in software development and leadership.
Josh Tobin
Founder @ GantryJosh Tobin is the founder and CEO of Gantry. Previously, Josh worked as a deep learning & robotics researcher at OpenAI and as a management consultant at McKinsey. He is also the creator of Full Stack Deep Learning (fullstackdeeplearning.com), the first course focused on the emerging engineering discipline of production machine learning. Josh did his PhD in Computer Science at UC Berkeley advised by Pieter Abbeel.
Sohini Roy
Senior Developer Relations Manager @ NVIDIASohini Bianka Roy is a senior developer relations manager at NVIDIA, working within the Enterprise Product group. With a passion for the intersection of machine learning and operations, Sohini specializes in the domains of MLOps and LLMOps. With her extensive experience in the field, she plays a crucial role in bridging the gap between developers and enterprise customers, ensuring smooth integration and deployment of NVIDIA's cutting-edge technologies. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/ Connect with Amrutha on LinkedIn: https://www.linkedin.com/in/amruthagujjar/ Connect with Josh on LinkedIn: https://www.linkedin.com/in/josh-tobin-4b3b10a9/ Connect with Sohini on Twitter: https://twitter.com/biankaroy_
MLOps Coffee Sessions #172 with Lingjiao Chen, FrugalGPT: Better Quality and Lower Cost for LLM Applications.
This episode is sponsored by QuantumBlack. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract There is a rapidly growing number of large language models (LLMs) that users can query for a fee. We review the cost associated with querying popular LLM APIs, e.g. GPT-4, ChatGPT, J1-Jumbo, and find that these models have heterogeneous pricing structures, with fees that can differ by two orders of magnitude. In particular, using LLMs on large collections of queries and text can be expensive. Motivated by this, we outline and discuss three types of strategies that users can exploit to reduce the inference cost associated with using LLMs: 1) prompt adaptation, 2) LLM approximation, and 3) LLM cascade. As an example, we propose FrugalGPT, a simple yet flexible instantiation of LLM cascade which learns which combinations of LLMs to use for different queries in order to reduce cost and improve accuracy. Our experiments show that FrugalGPT can match the performance of the best individual LLM (e.g. GPT-4) with up to 98% cost reduction or improve the accuracy over GPT-4 by 4% with the same cost. The ideas and findings presented here lay a foundation for using LLMs sustainably and efficiently. // Bio Lingjiao Chen is a Ph.D. candidate in the computer sciences department at Stanford University. He is broadly interested in machine learning, data management, and optimization. Working with Matei Zaharia and James Zou, he is currently exploring the fast-growing marketplaces of artificial intelligence and data. His work has been published at premier conferences and journals such as ICML, NeurIPS, SIGMOD, and PVLDB, and partially supported by a Google fellowship. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://lchen001.github.io/ FrugalGPT: How to Use Large Language Models While Reducing Cost and Improving Performance paper: https://arxiv.org/abs/2305.05176 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Lingjiao on LinkedIn: Timestamps: [00:00] Lingjiao's preferred coffee [00:35] Takeaways [02:41] Sponsor Ad: Nayur Khan of QuantumBlack [05:27] Lingjiao's research at Stanford [07:51] Day-to-day research overview [10:11] Inventing data management inspired abstractions research [13:58] Agnostic Approach to Data Management [15:56] Frugal GPT [18:59] Just another data provider [19:51] Frugal GPT breakdown [26:33] First step of optimizing the prompts [28:04] Prompt overlap [29:06] Query Concatenation [32:30] Money saving [35:04] Economizing the prompts [38:52] Questions to accommodate [41:33] LLM Cascade [47:25] Frugal GPT saves cost and Improves performance [51:37] End-user implementation [52:31] Completion Cache [56:33] Using a vector store [1:00:51] Wrap up
MLOps Coffee Sessions #172 with LLMs in Production Conference part 2 Building LLM Products Panel, George Mathew, Asmitha Rathis, Natalia Burina, and Sahar Mor Using hosted by TWIML's Sam Charrington. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract There are key areas we must be aware of when working with LLMs. High costs and low latency requirements are just the tip of the iceberg. In this panel, we hear about common pitfalls and challenges we must keep in mind when building on top of LLMs. // Bio Sam Charrington Sam is a noted ML/AI industry analyst, advisor and commentator, and host of the popular TWIML AI Podcast (formerly This Week in Machine Learning and AI). The show is one of the most popular Tech podcasts with nearly 15 million downloads. Sam has interviewed over 600 of the industry’s leading machine learning and AI experts and has conducted extensive research into enterprise AI adoption, MLOps, and other ML/AI-enabling technologies. George Mathew George is a Managing Director at Insight Partners focused on venture-stage investments in AI, ML, Analytics, and Data companies as they are establishing product/market Fit. Asmitha Rathis Asmitha is a Machine Learning Engineer with experience in developing and deploying ML models in production. She is currently working at an early-stage startup, PromptOps, where she is building conversational AI systems to assist developers. Prior to her current role, she was an ML engineer at VMware. Asmitha holds a Master’s degree in Computer Science from the University of California, San Diego, with a specialization in Machine Learning and Artificial Intelligence. Natalia Burina Natalia is an AI Product Leader who was most recently at Meta, leading Responsible AI. During her time at Meta, she led teams working on algorithmic transparency and AI Privacy. In 2017 Natalia was recognized by Business Insider as “The Most Powerful Female Engineer in 2017”. Natalia was also an Entrepreneur in Residence at Foundation Capital, advising portfolio companies and working with partners on deal flow. Prior to this, she was the Director of Product for Machine Learning at Salesforce, where she led teams building a set of AI capabilities and platform services. Prior to Facebook and Salesforce, Natalia led product development at Samsung, eBay, and Microsoft. She was also the Founder and CEO of Parable, a creative photo network bought by Samsung in 2015. Natalia started her career as a software engineer after pursuing Bachelor's degree in Applied and Computational Mathematics from the University of Washington. Sahar Mor Sahar is a Product Lead at Stripe with 15y of experience in product and engineering roles. At Stripe, he leads the adoption of LLMs and the Enhanced Issuer Network - a set of data partnerships with top banks to reduce payment fraud. Prior to Stripe he founded a document intelligence API company, was a founding PM in a couple of AI startups, including an accounting automation startup (Zeitgold, acq'd by Deel), and served in the elite intelligence unit 8200 in engineering roles. Sahar authors a weekly AI newsletter (AI Tidbits) and maintains a few open-source AI-related libraries (https://github.com/saharmor). // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/
MLOps Coffee Sessions #171 with Thibaut Labarre, Using Large Language Models at AngelList co-hosted by Ryan Russon. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract Thibaut innovatively addressed previous system constraints, achieving scalability and cost efficiency. Leveraging AngelList investing and natural language processing expertise, they refined news article classification for investor dashboards. Central is their groundbreaking platform, AngelList Relay, automating parsing and offering vital insights to investors. Amid challenges like Azure OpenAI collaboration and rate limit solutions, Thibaut reflects candidly. The narrative highlights prompt engineering's strategic importance and empowering domain experts for ongoing advancement. // Bio Thibaut LaBarre is an engineering lead with a background in Natural Language Processing (NLP). Currently, Thibaut focuses on unlocking the potential of Large Language Model (LLM) technology at AngelList, enabling everyone within the organization to become prompt engineers on a quest to streamline and automate the infrastructure for Venture Capital. Prior to that, Thibaut began his journey at Amazon as an intern where he built Heartbeat, a state-of-the-art NLP tool that consolidates millions of data points from various feedback sources, such as product reviews, customer contacts, and social media, to provide valuable insights to global product teams. Over the span of seven years, he expanded his internship project into an organization of 20 engineers. He received a M.S. in Computational Linguistics from the University of Washington. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://www.angellist.com/venture/relay --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ryan on LinkedIn: https://www.linkedin.com/in/ryanrusson/ Connect with Thibaut on LinkedIn: https://www.linkedin.com/in/thibautlabarre/
Timestamps: [00:00] Thibaut's preferred beverage [00:50] Takeaways [04:05] Please like, share, and subscribe to our MLOps channels! [04:44] A huge fan of Isaac Asimov [07:20] Thibaut Labarre background [09:13] AngelList as an organization [10:50] AI sense of building [12:29] System trade-offs [15:20] OpenAI's limitation [16:31] Human in the loop [17:22] Classifying relevance [18:09] Fight for value [19:37] Added value [22:10] Exploring efficient ways to automate tasks. [24:20] Investing in off-the-shelf models [27:56] AngelList Relay [30:49] News article and investment document classification technology [32:39] Back-end tech [34:09] Prompt layer [35:28] Prompt layer as a living [37:04] Foreseeing no human intervention [39:00] Blocking hallucinations [40:33] Challenges [43:49] Investments in other models besides OpenAI [45:20] Integration with other models [46:28] Ethical concerns when [48:37] OpenAI breaking Prompts [50:46] Wrap up
MLOps Coffee Sessions #170 with Phillip Carter, All the Hard Stuff with LLMs in Product Development.
We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O // Abstract
Delve into challenges in implementing LLMs, such as security concerns and collaborative measures against attacks. Emphasize the role of ML engineers and product managers in successful implementation. Explore identifying leading indicators and measuring ROI for impactful AI initiatives. // Bio Phillip is on the product team at Honeycomb where he works on a bunch of different developer tooling things. He's an OpenTelemetry maintainer -- chances are if you've read the docs to learn how to use OTel, you've read his words. He's also Honeycomb's (accidental) prompt engineering expert by virtue of building and shipping products that use LLMs. In a past life, he worked on developer tools at Microsoft, helping bring the first cross-platform version of .NET into the world and grow to 5 million active developers. When not doing computer stuff, you'll find Phillip in the mountains riding a snowboard or backpacking in the Cascades. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: https://phillipcarter.dev/ https://www.honeycomb.io/blog/improving-llms-production-observability https://www.honeycomb.io/blog/hard-stuff-nobody-talks-about-llm https://phillipcarter.dev/posts/how-to-make-an-fsharp-code-fixer/ The "hard stuff" post: https://www.honeycomb.io/blog/hard-stuff-nobody-talks-about-llm Our follow-up on iterating on LLMs in prod: https://www.honeycomb.io/blog/improving-llms-production-observability --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Phillip on LinkedIn: https://www.linkedin.com/in/phillip-carter-4714a135/ Timestamps: [00:00] Phillip's preferred coffee [00:33] Takeaways [01:53] Please like, share, and subscribe to our MLOps channels! [02:45] Phillip's background [07:15] Querying Natural Language [11:25] Function calls [14:29] Pasting errors or traces [16:30] Error patterns [20:22] Honeycomb's Improvement cycle [23:20] Prompt boxes rationale [28:06] Prompt injection cycles [32:11] Injection Attempt [33:30] UI undervalued, charging the AI feature [35:11] ROI cost [44:26] Bridging ML and Product Perspective [52:53] AI Model Trade-offs [56:33] Query assistant [59:07] Honeycomb is hiring! [1:00:08] Wrap up
MLOps Coffee Sessions #169 with Barak Turovsky, MLOps at the Age of Generative AI.
Thanks to wandb.ai for sponsoring this episode. Check out their new course on evaluating and fine-tuning LLMs wandb.me/genai-mlops.course // Abstract The talk focuses on MLOps aspects of developing, training and serving Generative AI/Large Language models // Bio Barak is an Executive in Residence at Scale Venture Partners, a leading Enterprise venture capital firm. Barak spent 10 years as Head of Product and User Experience for Languages AI and Google Translate teams within the Google AI org, focusing on applying cutting-edge Artificial Intelligence and Machine Learning technologies to deliver magical experiences across Google Search, Assistant, Cloud, Chrome, Ads, and other products. Previously, Barak spent 2 years as a product leader within the Google Commerce team. Most recently, Barak served as Chief Product Officer, responsible for product management and engineering at Trax, a leading provider of Computer Vision AI solutions for Retail and Commerce industries. Prior to joining Google in 2011, Barak was Director of Products in Microsoft’s Mobile Advertising, Head of Mobile Commerce at PayPal, and Chief Technical Officer at an Israeli start-up. He lived more than 10 years in 3 different countries (Russia, Israel, and the US) and fluently speaks three languages. Barak earned a Bachelor of Laws degree from Tel Aviv University, Israel, and a Master’s of Business Administration from the University of California, Berkeley. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Bio and links about Barak's work: https://docs.google.com/document/d/1E4Yrmt_Y57oTEYHQQDvt71XzSJ8Ew5WvscAQbHV4K3U/edit Framework for evaluating Generative AI use cases: https://www.linkedin.com/pulse/framework-evaluating-generative-ai-use-cases-barak-turovsky/?trackingId=%2BMRxEZ9WTPCNH2JscILTeg%3D%3D The Great A.I. Awakening: https://www.nytimes.com/2016/12/14/magazine/the-great-ai-awakening.html --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Barak on LinkedIn: https://www.linkedin.com/in/baraktur/ Timestamps: [00:00] Barak's preferred coffee [00:23] Barak Turovsky's background [03:10] Please like, share, and subscribe to our MLOps channels! [04:09] Getting into tech [08:39] First wave of AI [12:39] Building a product at a scale and the challenges [15:59] Framework for evaluating Generative AI use cases [24:33] Machine trust adoption [29:13] Wandb's new course [31:10] Focus on achievable use cases for LLMs. [36:36] User feedback [38:23] Disruption of entertainment and customer interactions [46:14] Get new tools or work with your own distribution? [47:57] Importance of data engineers [53:28] ML Engineers Collaborate with Product [56:13] Wrap up
MLOps Coffee Sessions #168 with Piotr Niedźwiedź, Experiment Tracking in the Age of LLMs, co-hosted by Vishnu Rachakonda.
// Abstract
Piotr shares his journey as an entrepreneur and the importance of focusing on core values to achieve success. He highlights the mission of Neptune to support ML teams by providing them with control and confidence in their models. The conversation delves into the role of experiment tracking in understanding and debugging models, comparing experiments, and versioning models. Piotr introduces the concept of prompt engineering as a different approach to building models, emphasizing the need for prompt validation and testing methods.
// Bio
Piotr is the CEO of neptune.ai. Day to day, apart from running the company, he focuses on the product side of things. Strategy, planning, ideation, getting deep into user needs and use cases. He really likes it.
Piotr's path to ML started with software engineering. Always liked math and started programming when he was 7. In high school, Piotr got into algorithmics and programming competitions and loved competing with the best. That got him into the best CS and Maths program in Poland which funny enough today specializes in machine learning.
Piotr did his internships at Facebook and Google and was offered to stay in the Valley. But something about being a FAANG engineer didn’t feel right. He had this spark to do more, build something himself. So with a few of his friends from the algo days, they started Codilime, a software consultancy, and later a sister company Deepsense.ai machine learning consultancy, where he was a CTO.
When he came to the ML space from software engineering, he was surprised by the messy experimentation practices, lack of control over model building, and a missing ecosystem of tools to help people deliver models confidently.
It was a stark contrast to the software development ecosystem, where you have mature tools for DevOps, observability, or orchestration to execute efficiently in production. And then, one day, some ML engineers from Deepsense.ai came to him and showed him this tool for tracking experiments they built during a Kaggle competition (which they won btw), and he knew this could be big.
He asked around, and everyone was struggling with managing experiments. He decided to spin it off as a VC-funded product company, and the rest is history.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://neptune.ai/blog/author/piotr-niedzwiedz
https://www.youtube.com/playlist?list=PLKePQLVx9tOfKFbg9GY2Anl41Be4T1-m5
https://thesequence.substack.com/p/-piotr-niedzwiedz-neptunes-ceo-on
https://open.spotify.com/episode/2KEqTMAHODbPKdUEtlrhm7?si=ed862b2ac7534e39
https://www.linkedin.com/in/piotrniedzwiedz/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Piotr on LinkedIn: https://www.linkedin.com/in/piotrniedzwiedz/
Timestamps:
[00:00] Introduction to Piotr Niedźwiedź
[01:35] Please like, share, and subscribe to our MLOps channels!
[01:58] Wojciech Zaremba
[05:20] The Olympiad
[06:31] Building own company
[12:16] Talking outside Poland with the same passion
[13:45] Adapting with Neptune
[19:35] Core values focus
[22:02] Product Vision changes with advances
[29:36] Control and confidence
[30:05] Experiment tracking existing use cases
[37:25] Control pane
[38:59] Piotr's prediction
[43:20] WiFi issues around the world
[44:09] Wrap up
MLOps Coffee Sessions #167 with Maxime Beauchemin, Treating Prompt Engineering More Like Code. // Abstract Promptimize is an innovative tool designed to scientifically evaluate the effectiveness of prompts. Discover the advantages of open-sourcing the tool and its relevance, drawing parallels with test suites in software engineering. Uncover the increasing interest in this domain and the necessity for transparent interactions with language models. Delve into the world of prompt optimization, deterministic evaluation, and the unique challenges in AI prompt engineering. // Bio Maxime Beauchemin is the founder and CEO of Preset, a series B startup supporting and commercializing the Apache Superset project. Max was the original creator of Apache Airflow and Apache Superset when he was at Airbnb. Max has over a decade of experience in data engineering, at companies like Lyft, Airbnb, Facebook, and Ubisoft. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Max's first MLOps Podcast episode: https://go.mlops.community/KBnOgN Test-Driven Prompt Engineering for LLMs with Promptimize blog: https://maximebeauchemin.medium.com/mastering-ai-powered-product-development-introducing-promptimize-for-test-driven-prompt-bffbbca91535https://maximebeauchemin.medium.com/mastering-ai-powered-product-development-Test-Driven Prompt Engineering for LLMs with Promptimize podcast: https://talkpython.fm/episodes/show/417/test-driven-prompt-engineering-for-llms-with-promptimizeTaming AI Product Development Through Test-driven Prompt Engineering // Maxime Beauchemin // LLMs in Production Conference lightning talk: https://home.mlops.community/home/videos/taming-ai-product-development-through-test-driven-prompt-engineering --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Max on LinkedIn: https://www.linkedin.com/in/maximebeauchemin/ Timestamps: [00:00] Max introducing the Apache Superset project at Preset [01:04] Max's preferred coffee [01:16] Airflow creator [01:45] Takeaways [03:53] Please like, share, and subscribe to our MLOps channels! [04:31] Check Max's first MLOps Podcast episode [05:20] Promptimize [06:10] Interaction with API [08:27] Deterministic evaluation of SQL queries and AI [12:40] Figuring out the right edge cases [14:17] Reaction with Vector Database [15:55] Promptomize Test Suite [18:48] Promptimize vision [20:47] The open-source blood [23:04] Impact of open source [23:18] Dangers of open source [25:25] AI-Language Models Revolution [27:36] Test-driven design [29:46] Prompt tracking [33:41] Building Test Suites as Assets [36:49] Adding new prompt cases to new capabilities [39:32] Monitoring speed and cost [44:07] Creating own benchmarks [46:19] AI feature adding more value to the end users [49:39] Perceived value of the feature [50:53] LLMs costs [52:15] Specialized model versus Generalized model [56:58] Fine-tuning LLMs use cases [1:02:30] Classic Engineer's Dilemma [1:03:46] Build exciting tech that's available [1:05:02] Catastrophic forgetting [1:10:28] Promt driven development [1:13:23] Wrap up
MLOps Coffee Sessions #166 with Roy Hasson & Santona Tuli, Eliminating Garbage In/Garbage Out for Analytics and ML. // Abstract Shift left data quality ownership and observability that makes it easy for users to catch bad data at the source and stop it from entering your analytics/ML stack. // Bio Santona Tuli Santona Tuli, Ph.D. began her data journey through fundamental physics—searching through massive event data from particle collisions at CERN to detect rare particles. She’s since extended her machine learning engineering to natural language processing, before switching focus to product and data engineering for data workflow authoring frameworks. As a Python engineer, she started with the programmatic data orchestration tool, Airflow, helping improve its developer experience for data science and machine learning pipelines. Currently, at Upsolver, she leads data engineering and science, driving developer research and engagement for the declarative workflow authoring framework in SQL. Dr. Tuli is passionate about building, as well as empowering others to build, end-to-end data and ML pipelines, scalably. Roy Hasson Roy is the head of product at Upsolver helping companies deliver high-quality data to their analytics and ML tools. Previously, Roy led product management for AWS Glue and AWS Lake Formation. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://royondata.substack.com/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Roy on LinkedIn: https://www.linkedin.com/in/royhasson/ Connect with Santona on LinkedIn: https://www.linkedin.com/in/santona-tuli/ Timestamps: [00:00] Santona's and Roy's preferred coffee [01:05] Santona's and Roy's background [03:33] Takeaways [05:49] Please like, share, and subscribe to our MLOps channels! [06:42] Back story of having Santona and Roy on the podcast [09:51] Santona's story [11:37] Optimal tag teamwork [16:53] Dealing with stakeholder needs [26:25] Having mechanisms in place [27:30] Building for data Engineers vs building for data scientists [34:50] Creating solutions for users [38:55] User experience holistic point of view [41:11] Tooling sprawl is real [42:00] LLMs reliability [45:00] Things would have loved to learn five years ago [49:46] Wrap up
MLOps Coffee Sessions #165 with Sammy Sidhu, Python Power: How Daft Embeds Models and Revolutionizes Data Processing. // Abstract Sammy shares his fascinating journey in the autonomous vehicle industry, highlighting his involvement in two successful startup acquisitions by Tesla and Toyota. He emphasizes his expertise in optimizing and distilling models for efficient machine learning, which he has incorporated into his new company Eventual. The company's open-source offering, daf, focuses on tackling the challenges of unstructured and complex data. Sammy discusses the future of MLOps, machine learning, and data storage, particularly in relation to the retrieval and processing of unstructured data. The Eventual team is developing Daft, an open-source query engine that aims to provide efficient data storage solutions for unstructured data, offering features like governance, schema evolution, and time travel. The conversation sheds light on the innovative developments in the field and the potential impact on various industries. // Bio Sammy is a Deep Learning and systems veteran, holding over a dozen publications and patents in the space. Sammy graduated from the University of California, Berkeley where he did research in Deep Learning and High Performance Computing. He then joined DeepScale as the Chief Architect and led the development of perception technologies for autonomous vehicles. During this time, DeepScale grew rapidly and was subsequently acquired by Tesla in 2019. Staying in Autonomous Vehicles, Sammy joined Lyft Level 5 as a Senior Staff Software Engineer, building out core perception algorithms as well as infrastructure for machine learning and embedded systems. Level 5 was then acquired by Toyota in 2021, adopting much of his work. Sammy is now CEO and Co-Founder at Eventual Building Daft, an open-source query engine that specializes in multimodal data. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://sammysidhu.com/ Check out Daft, our open-source query engine for multimodal data! https://www.getdaft.io/ Here are some talks/shows we have given about it: - PyData Global (Dec 2022): Large-scale image processing: https://www.youtube.com/watch?v=ol6IQUbyeDo&ab_channel=PyData - Ray Meetup (March 2023): Distributed ML preprocessing + training on Ray https://www.youtube.com/watch?v=1MpEYlIlu7w&t=2972s&ab_channel=Anyscale - The Data Stack Show (April 2023): Self-Driving Technology and Data Infrastructure with Sammy Sidhu https://datastackshow.com/podcast/the-prql-self-driving-technology-and-data-infrastructure-with-sammy-sidhu-co-founder-and-ceo-of-eventual/ Chain of Thought for LLMs: https://cobusgreyling.medium.com/chain-of-thought-prompting-in-llms-1077164edf97 Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes: https://arxiv.org/abs/2305.02301 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Sammy on LinkedIn: https://www.linkedin.com/in/sammy-sidhu/
MLOps Coffee Sessions #164 with Rob Hirschfeld, Open Source and Fast Decision Making. This episode is brought to you by. // Abstract Rob Hirschfeld, the CEO and co-founder of Rack N, discusses his extensive experience in the DevOps movement. He shares his notable achievement of coining the term "the cloud" and obtaining patents for infrastructure management and API provision. Rob highlights the stagnant progress in operations and the persistent challenges in security and access controls within the industry. The absence of standardization in areas such as Kubernetes and single sign-on complicates the development of robust solutions. To address these issues, Rob underscores the significance of open-source practices, automation, and version control in achieving operational independence and resilience in infrastructure management. // Bio Rob is the CEO and Co-founder of RackN, an Austin-based start-up that develops software to help automate data centers, which they call Digital Rebar. This platform helps connect all the different pieces and tools that people use to manage infrastructure into workflow pipelines through seamless multi-component automation across the different pieces and parts needed to bring up IT systems, platforms, and applications. Rob has a background in Scale Computing, Mechanical and Systems Engineering, and specializes in large-scale complex systems that are integrated with the physical environment. He has founded companies and been in the cloud and infrastructure space for nearly 25 years and has done everything from building the first Clouds using ESXi betas to serving four terms on the OpenStack Foundation Board. Rob was trained as an Industrial Engineer and holds degrees from Duke University and Louisiana State University. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://rackn.com/ https://robhirschfeld.com/about/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Rob on LinkedIn: https://www.linkedin.com/in/rhirschfeld/ Timestamps: [00:00] Rob's preferred coffee [00:17] Rob Hirschfeld's background [01:42] Takeaways [02:36] Please like, share, and subscribe to this channel! [03:09] Creation of Cloud [08:38] Changes in Cloud after 25 Years [10:54] Pros and cons of microservices [13:06] Secure Access Provisioning [15:46] Parallelism with ads [18:08] Redfish protocol [20:21] Impact of using open source vs using a SAS provider [26:15] Automation [32:39] Embrace Operational Flexibility [35:08] Automating infrastructure inefficiently [41:26] Legacy code and resiliency [43:39] Collection of metadata [45:50] RackN [51:23] Granular Cloud Preferences [54:35] Reframing of perceived complexity [57:32] Generative DevOps [58:50] Wrap up
MLOps Coffee Sessions #163 with Yujian Tang, Democratizing AI co-hosted by Abi Aryan. // Abstract The popularity of ChatGPT has brought large language model (LLM) apps and their supporting technologies to the forefront. One of the supporting technologies is vector databases. Yujian shares how vector databases like Milvus are used in production and how they solve one of the biggest problems in LLM app building - data issues. They also discuss how Zilliz is democratizing vector databases through education, expanding access to technologies, and technical evangelism. // Bio Yujian Tang is a Developer Advocate at Zilliz. He has a background as a software engineer working on AutoML at Amazon. Yujian studied Computer Science, Statistics, and Neuroscience with research papers published to conferences including IEEE Big Data. He enjoys drinking bubble tea, spending time with family, and being near water. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Yujian on LinkedIn: https://www.linkedin.com/in/yujiantang Timestamps: [00:00] Yujian's preferred coffee [02:40] Takeaways [05:14] Please share this episode with your friends! [06:39] Vector databases trajectory [09:00] 2 start-up companies created by Yujian [09:39] Uninitiated Vector Databases [12:20] Vector Databases trade-off [14:16] Difficulties in training LLMs [23:30] Enterprise use cases [27:38] Process/rules not to use LLMs unless necessary [32:14] Setting up returns [33:13] When not to use Vector Databases [35:30] Elastic search [36:07] Generative AI apps common pitfalls [39:35] Knowing your data [41:50] Milvus [48:28] Actual Enterprise use cases [49:32] Horror stories [50:31] Data mesh [51:06] GPTCash [52:10] Shout out to the Seattle Community! [53:44] Wrap up
MLOps Coffee Sessions #162 with Soham Chatterjee, From LLMs to TinyML: The Dynamic Spectrum of MLOps co-hosted by Abi Aryan. // Abstract Explore the spectrum of MLOps from large language models (LLMs) to TinyML. Soham highlights the difficulties of scaling machine learning models and cautions against relying exclusively on open AI's API due to its limitations. Soham is particularly interested in the effective deployment of models and the integration of IoT with deep learning. He offers insights into the challenges and strategies involved in deploying models in constrained environments, such as remote areas with limited power and utilizing small devices like Arduino Nano. // Bio Soham leads the machine learning team at Sleek, where he builds tools for automated accounting and back-office management. As an electrical engineer, Soham has a passion for the intersection of machine learning and electronics, specifically TinyML/Edge Computing. He has several courses on MLOps and TinyMLOps available on Udacity and LinkedIn, with more courses in the works. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Soham on LinkedIn: https://www.linkedin.com/in/soham-chatterjee Timestamps: [00:00] Soham's preferred coffee [01:49] Takeaways [05:33] Please share this episode with [07:02] Soham's background [09:00] From electrical engineering to Machine Learning [10:40] Deep learning, Edge Computing, and Quantum Computing [11:34] Tiny ML [13:29] Favorite area in Tiny ML chain [14:03] Applications explored [16:56] Operational challenges transformation [18:49] Building with Large Language Models [25:44] Most Optimal Model [26:33] LLMs path [29:19] Prompt engineering [33:17] Migrating infrastructures to new product [37:20] Your success where others failed [38:26] API Accessibility [39:02] Reality about LLMs [40:39] Compression angle adds to the bias [43:28] Wrap up
MLOps Coffee Sessions #161 with Tuhin Srivastava, The Long Tail of ML Deployment co-hosted by Abi Aryan. This episode is brought to you by QuantumBlack. // Abstract Baseten is an engineer-first platform designed to alleviate the engineering burden for machine learning and data engineers. Tuhin's perspective, based on research with Stanford students, emphasizes the importance of engineers embracing the engineering aspects and considering them from a reproductive perspective. // Bio Tuhin Srivastava is the co-founder and CEO of Baseten. Tuhin has spent the better part of the last decade building machine learning-powered products and is currently working on empowering engineers to build production-grade services with machine learning. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links QuantumBlack: https://www.mckinsey.com/capabilities/quantumblack/contact-us
Baseten: https://www.baseten.co/ Baseten Careers: https://www.baseten.co/careers --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Tuhin on LinkedIn: https://www.linkedin.com/in/tuhin-srivastava-60601114/ Timestamps: [00:00] Partnership with QuantumBlack [00:16] Nayur Khan presenting QuantumBlack [03:35] QuantumBlack is hiring! [03:47] Tuhin's preferred coffee [05:03] Takeaways [07:00] Please share this episode with a friend! [07:12] Comments/Reviews [08:49] Tuhin's background [10:13] Finance and Law common complaint culture [11:40] Doing Machine Learning in 2010 - 2011 [14:31] Gum broad or the next company shape? [16:33] Engineers need to learn machine learning [20:18] Software engineers need to dig deeper [24:49] Cambrian Explosion [27:53] The Holy Trifecta [28:54] Objective truth and prompting [31:23] Limitations of LLMs [35:37] Documentation challenges [38:25] Baseten creating valuable models [40:37] Advocate for Microservices or API-based solution [42:54] Learning Git pains [44:16] Baseten back ups [48:00] Baseten is hiring! [49:32] Wrap up
MLOps Coffee Sessions #160 with Matt Sharp, Data Developer at Shopify, Clean Code for Data Scientists co-hosted by Abi Aryan. // Abstract Let's delve into Shopify's real-time serving platform, Merlin, which enables features like recommender systems, inbox classification, and fraud detection. Matt shares his insights on clean coding and the new book he is writing about LLMs in production. // Bio Matt is a Chemical Engineer turned Data scientist turned Data Engineer. Self-described "Recovering Data Scientist", Matt got tired of all the inefficiencies he faced as a Data Scientist and made the switch to Data Engineering. At Matt's last job, he ended up building the entire MLOps platform from scratch for a fintech startup called MX. Matt gives tips to data scientists on LinkedIn on how to level up their careers and has started to be known for my clean code tips in particular. Matt recently just started a new job at Shopify. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Matt on LinkedIn: https://www.linkedin.com/in/matthewsharp/
MLOps Coffee Sessions #159 with Maria Vechtomova, Lead ML engineer, and Basak Eskili Machine Learning Engineer, at Ahold Delhaize, Why is MLOps Hard in an Enterprise? co-hosted by Abi Aryan. // Abstract MLOps is particularly challenging to implement in enterprise organizations due to the complexity of the data ecosystem, the need for collaboration across multiple teams, and the lack of standardization in ML tooling and infrastructure. In addition to these challenges, at Ahold Delhaize, there is a requirement for the reusability of models as our brands seek to have similar data science products, such as personalized offers, demand forecasts, and cross-sell. // Bio Maria Vechtomova Maria is a Machine Learning Engineer at Ahold Delhaize. Maria is bridging the gap between data scientists infra and IT teams at different brands and focuses on standardization of machine learning operations across all the brands within Ahold Delhaize. During nine years in Data&Analytics, Maria tried herself in different roles, from data scientist to machine learning engineer, was part of teams in various domains, and has built broad knowledge. Maria believes that a model only starts living when it is in production. For this reason, last six years, her focus was on the automation and standardization of processes related to machine learning. Basak Eskili Basak Eskili is a Machine Learning Engineer at Ahold Delhaize. She is working on creating new tools and infrastructure that enable data scientists to quickly operationalise algorithms. She is bridging the space between data scientists and platform engineers while improving the way of working in accordance with MLOps principles. In her previous role, she was responsible for bringing models to production. She focused on NLP projects and building data processing pipelines. Basak also implemented new solutions by using cloud services for existing applications and databases to improve time and efficiency. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links MLOps Maturity Assessment Blog: https://mlops.community/mlops-maturity-assessment/ The Minimum Set of Must-Haves for MLOps Blog: https://mlops.community/the-minimum-set-of-must-haves-for-mlops/ Traceability & Reproducibility Blog: https://mlops.community/traceability-reproducibility/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Maria on LinkedIn: https://www.linkedin.com/in/maria-vechtomova/Connect with Basak on LinkedIn: https://www.linkedin.com/in/ba%C5%9Fak-tu%C4%9F%C3%A7e-eskili-61511b58/
MLOps Coffee Sessions #158 with Nils Reimer, MLOps Build or Buy, Large Language Model at Scale co-hosted by Abi Aryan. // Abstract Large Language Models with billions of parameters have the possibility to change how we work with textual data. However, running them on scale at potentially hundred millions of texts a day is a massive challenge. Nils talks about finding the right model size for respective tasks, model distillation, and promising new ways on transferring knowledge from large to smaller models. // Bio Nils Reimers is highly recognized throughout the AI community for creating and maintaining the now-famous Sentence Transformers library (www.SBERT.net) used to develop, train, and use state-of-the-art LLMs. The project has 900+ stars on GitHub and 30M+ installations. Nils is currently the Director of Machine Learning at Cohere where he leads the team that develops and trains Large Language Models (LLM) with billions of parameters. Prior to Cohere, Nils created and led the science team for Neural Search at HuggingFace. Nils holds a Ph.D. in Computer Science from UKP in Darmstadt. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links (www.SBERT.net) https://www.nils-reimers.de/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Nils on LinkedIn: https://www.linkedin.com/in/reimersnils/ Timestamps: [00:00] Nils' preferred coffee [00:45] Nils' background [01:30] Takeaways [06:47] Subscribe to our Newsletters and IRL Meetups, and leave your reviews! [07:32] Nils' history [10:39] From IT Security to Machine Learning [13:22] Tangibility of IT and Security [14:46] NLP transition [15:55] Bad augmentation to new capabilities of LLMs [16:59] Nils' concern during his PH.D. [19:55] Making Money from Machine Learning [22:06] Catastrophic forgetting [26:34] Updating solutions [28:42] Neural search space and building adaptive models [31:23] Filtering models [32:32] Latency issues [36:53] Models running in parallel [37:54] Generative models problems [38:43] Nils' role at Cohere [41:41] To build or not to build API [43:00] Search models [45:38] Large use cases [46:43] Open source discussion within Cohere [50:48] Competitive Edge [55:27] Future world of API [58:14] LLMs in Production Conference Part 2 announcement! [1:00:17] Hopeful direction of Cohere's future [1:02:33] Data silos [1:04:34] Where to update the database and code [1:05:24] Nils' focus [1:08:49] Make money or save money [1:10:30] Cohere's upcoming project [1:12:37] Time spent red teaming the models [1:14:05] Wrap up
We are having another LLMs in-production Virtual Conference. 50+ speakers combined with in-person activities around the world on June 15 & 16. Sign up free here: https://home.mlops.community/home/events/llm-in-prod-part-ii-2023-06-20 // Abstract This panel discussion is centered around a crucial topic in the tech industry - data privacy and security in the context of large language models and AI systems. The discussion highlights several key themes, such as the significance of trust in AI systems, the potential risks of hallucinations, and the differences between low and high-affordability use cases. The discussion promises to be thought-provoking and informative, shedding light on the latest developments and concerns in the field. We can expect to gain valuable insights into an issue that is becoming increasingly relevant in our digital world. // Bio Diego Oppenheimer Diego Oppenheimer is an entrepreneur, product developer, and investor with an extensive background in all things data. Currently, he is a Partner at Factory a venture fund specializing in AI investments as well as interim head of product at two LLM startups. Previously he was an executive vice president at DataRobot, Founder, and CEO at Algorithmia (acquired by DataRobot), and shipped some of Microsoft’s most used data analysis products including Excel, PowerBI, and SQL Server. Diego is active in AI/ML communities as a founding member and strategic advisor for the AI Infrastructure Alliance and MLops.Community and works with leaders to define ML industry standards and best practices. Diego holds a Bachelor's degree in Information Systems and a Masters degree in Business Intelligence and Data Analytics from Carnegie Mellon University Gevorg Karapetyan Gevorg Karapetyan is the co-founder and CTO of ZERO Systems where he oversees the company's product and technology strategy. He holds a Ph.D. in Computer Science and is the author of multiple publications, including a US Patent. Vin Vashishta C-level Technical Strategy Advisor and Founder of V Squared, one of the first data science consulting firms. Our mission is to provide support and clarity for our clients’ complete data and AI monetization journeys. Over a decade in data science and a quarter century in technology building and leading teams and delivering products with $100M+ in ARR. Saahil Jain Saahil Jain is an engineering manager at You.com. At You.com, Saahil builds search, ranking, and conversational AI systems. Previously, Saahil was a graduate researcher in the Stanford Machine Learning Group under Professor Andrew Ng, where he researched topics related to deep learning and natural language processing (NLP) in resource-constrained domains like healthcare. Prior to Stanford, Saahil worked as a product manager at Microsoft on Office 365. He received his B.S. and M.S. in Computer Science at Columbia University and Stanford University respectively. Shreya Rajpal Shreya is the creator of Guardrails AI, an open-source solution designed to establish guardrails for large language models. As a founding engineer at Predibase, she helped build the Applied ML and ML infra teams. Previously, she worked at Apple's Special Projects Group on cross-functional ML, and at Drive.ai building computer vision models.
MLOps Coffee Sessions #157 with Katrina Ni & Aaron Maurer, MLOps Build or Buy, Startup vs. Enterprise? co-hosted by Jake Noble of tecton.ai.
This episode is sponsored by tecton.ai - Check out their feature store to get your real-time ML journey started. // Abstract There are a bunch of challenges with building useful machine learning at a B2B software company like Slack, but we've built some cool use cases over the years, particularly around recommendations. One of the key challenges is how to train powerful models while being prudent stewards of our clients' essential business data, and how to do so while respecting the increasingly complex landscape of international data regulation. // Bio Katrina Ni Katrina is a Machine Learning Engineer in Slack ML Services Team where they build ML platforms and integrate ML, e.g. Recommend API, Spam Detection, across product functionalities. Prior to Slack, she is a Software Engineer in Tableau Explain Data Team where they build tools that utilize statistical models and propose possible explanations to help users inspect, uncover, and dig deeper into the viz. Aaron Maurer Aaron is a senior engineering manager in the infra organization at Slack, managing both the machine learning team and the real-time services team. In six years at Slack, most of which Aaron spent as an engineer, He worked on the search ranking, recommendation, spam detection, performance anomaly detection, and many other ML applications. Aaron is also an advisor to Eppo, an experimentation platform. Prior to Slack, Aaroon worked as a data scientist at Airbnb, earned a Masters in statistics at the University of Chicago, and helped develop econometric models projecting the Obamacare rollout at Acumen LLC. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Jake on LinkedIn: https://www.linkedin.com/in/jakednoble/ Connect with Katrina on LinkedIn: https://www.linkedin.com/in/katrina-ni-660b2590/ Connect with Aaron on LinkedIn: https://www.linkedin.com/in/aaron-maurer-4003b638/ Timestamps: [00:00] Aaron and Katrina's preferred coffee [00:41] Recommender and System and Jake [02:06] Takeaways [05:38] Introduction to Aaron Maurer & Katrina Ni [06:53] Aaron Maurer & Katrina Ni's Recommend API blog post [08:36] 10-pole machine learning use case and Rex's use case [10:14] Genesis of Slack's recommender system framework [11:47] The Special Sauce [12:58] Speaking the same language [15:23] Use case sources [17:08] Slack's feature engineering [17:52] Main CTR models [18:40] Data privacy [21:33] Slack's recommendations problem [22:09] Fine-tuning the generative models [23:30] Cold start problem [26:02] Underrated [28:24] Baseline [28:55] Cold sore space [30:15] LLMs in Production Conference Part 2 announcement! [32:32] Data scientists transition to ML [33:35] Unicorns do exist! [34:43] Diversity of skill set [36:02] The future of ML [38:34] Model Serving [40:11] MLOps Maturity level [43:06] AWS Analogy [45:05] Primary difficulty [48:07] Wrap up
Sign up for the next LLM in production conference here: https://go.mlops.community/LLMinprod
Watch all the talks from the first conference: https://go.mlops.community/llmconfpart1
// Abstract In this panel discussion, the topic of the cost of running large language models (LLMs) is explored, along with potential solutions. The benefits of bringing LLMs in-house, such as latency optimization and greater control, are also discussed. The panelists explore methods such as structured pruning and knowledge distillation for optimizing LLMs. OctoML's platform is mentioned as a tool for the automatic deployment of custom models and for selecting the most appropriate hardware for them. Overall, the discussion provides insights into the challenges of managing LLMs and potential strategies for overcoming them. // Bio Lina Weichbrodt Lina is a pragmatic freelancer and machine learning consultant that likes to solve business problems end-to-end and make machine learning or a simple, fast heuristic work in the real world. In her spare time, Lina likes to exchange with other people on how they can implement best practices in machine learning, talk to her at the Machine Learning Ops Slack: shorturl.at/swxIN. Luis Ceze Luis Ceze is Co-Founder and CEO of OctoML, which enables businesses to seamlessly deploy ML models to production making the most out of the hardware. OctoML is backed by Tiger Global, Addition, Amplify Partners, and Madrona Venture Group. Ceze is the Lazowska Professor in the Paul G. Allen School of Computer Science and Engineering at the University of Washington, where he has taught for 15 years. Luis co-directs the Systems and Architectures for Machine Learning lab (sampl.ai), which co-authored Apache TVM, a leading open-source ML stack for performance and portability that is used in widely deployed AI applications. Luis is also co-director of the Molecular Information Systems Lab (misl.bio), which led pioneering research in the intersection of computing and biology for IT applications such as DNA data storage. His research has been featured prominently in the media including New York Times, Popular Science, MIT Technology Review, and the Wall Street Journal. Ceze is a Venture Partner at Madrona Venture Group and leads their technical advisory board. Jared Zoneraich Co-Founder of PromptLayer, enabling data-driven prompt engineering. Compulsive builder. Jersey native, with a brief stint in California (UC Berkeley '20) and now residing in NYC. Daniel Campos Hailing from Mexico Daniel started his NLP journey with his BS in CS from RPI. He then worked at Microsoft on Ranking at Bing with LLM(back when they had 2 commas) and helped build out popular datasets like MSMARCO and TREC Deep Learning. While at Microsoft he got his MS in Computational Linguistics from the University of Washington with a focus on Curriculum Learning for Language Models. Most recently, he has been pursuing his Ph.D. at the University of Illinois Urbana Champaign focusing on efficient inference for LLMs and robust dense retrieval. During his Ph.D., he worked for companies like Neural Magic, Walmart, Qualtrics, and Mendel.AI and now works on bringing LLMs to search at Neeva. Mario Kostelac Currently building AI-powered products in Intercom in a small, highly effective team. I roam between practical research and engineering but lean more towards engineering and challenges around running reliable, safe, and predictable ML systems. You can imagine how fun it is in LLM era :). Generally interested in the intersection of product and tech, and building a differentiation by solving hard challenges (technical or non-technical). Software engineer turned into Machine Learning engineer 5 years ago.
MLOps Coffee Sessions #156 with Melissa Barr & Michael Mui, Machine Learning Education at Uber co-hosted by Lina Weichbrodt. // Abstract Melissa and Michael discuss the education program they developed for Uber's machine learning platform service, Michelangelo, during a guest appearance on a podcast. The program teaches employees how to use machine learning both in general and specifically for Uber. The platform team can obtain valuable feedback from users and use it to enhance the platform. The course was designed using engineering principles, making it applicable to other products as well. // Bio Melissa Barr Melissa is a Technical Program Manager for ML & AI at Uber. She is based in New York City. She drives projects across Uber’s ML platform, delivery, and personalization teams. She also built out the first version of the ML Education Program in 2021. Michael Mui Melissa is a Staff Technical Lead Manager on Uber AI's Machine Learning Platform team. He leads the Distributed ML Training team which focuses on building elastic, scalable, and fault-tolerant distributed machine learning libraries and systems used to power machine learning development productivity across Uber. He also co-leads Uber’s internal ML Education initiatives. Outside of Uber, Michael also teaches ML at the Parsons School of Design in NYC as an Adjunct Faculty (mostly for the museum passes!) and guest lectures at the University of California, Berkeley. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://www.uber.com/blog/ml-education-at-uber-program-design-and-outcomes/https://www.uber.com/blog/ml-education-at-uber/https://www.uber.com/en-PH/blog/ml-education-at-uber/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Melissa on LinkedIn: https://www.linkedin.com/in/melissabarr1/ Connect with Michael on LinkedIn: https://www.linkedin.com/in/michael-c-mui/Connect with Lina on LinkedIn: https://www.linkedin.com/in/lina-weichbrodt-344a066a/
Timestamps: [00:00] Melissa and Michael's preferred coffee [01:51] Takeaways [05:40] Please subscribe to our newsletters and leave reviews on our podcasts! [06:18] Machine learning at Uber education program [07:45] The Uber courses [10:03] Tailoring the Uber education system [12:27] Growing out of the ML-Ed platform efforts [14:14] Expanding the ML Market Size [15:23] Relationship evolution [17:36] Reproducibility best practices [21:46] Learning development timeline [26:29] Courses effectiveness evaluation [29:57] Tracking Progress Challenge [31:25] ML platforms for internal tools [35:07] Impact of ML Education at Uber [39:30] Recommendations to companies who want to start an ML-Ed platform [41:12] Early ML Adoption Program [42:11] Homegrown or home-built platform [42:54] Feature creation to a course [45:24] ML Education at Uber: Frameworks Inspired by Engineering Principles [49:42] The Future of ML Education at Uber [52:28] Unclear ways to spread ML knowledge [54:20] Module for Generative AI and ChatGPT [55:05] Measurement of success [56:39] Wrap up
MLOps Coffee Sessions #155 with Matei Zaharia, The Birth and Growth of Spark: An Open Source Success Story, co-hosted by Vishnu Rachakonda. // Abstract We dive deep into the creation of Spark, with the creator himself - Matei Zaharia Chief technologist at Databricks. This episode also explores the development of Databricks' other open source home run ML Flow and the concept of "lake house ML". As a special treat Matei talked to us about the details of the "DSP" (Demonstrate Search Predict) project, which aims to enable building applications by combining LLMs and other text-returning systems. // About the guest: Matei has the unique advantage of being able to see different perspectives, having worked in both academia and the industry. He listens carefully to people's challenges and excitement about ML and uses this to come up with new ideas. As a member of Databricks, Matei also has the advantage of applying ML to Databricks' own internal practices. He is constantly asking the question "What's a better way to do this?" // Bio Matei Zaharia is an Associate Professor of Computer Science at Stanford and Chief Technologist at Databricks. He started the Apache Spark project during his Ph.D. at UC Berkeley, and co-developed other widely used open-source projects, including MLflow and Delta Lake, at Databricks. At Stanford, he works on distributed systems, NLP, and information retrieval, building programming models that can combine language models and external services to perform complex tasks. Matei’s research work was recognized through the 2014 ACM Doctoral Dissertation Award for the best Ph.D. dissertation in computer science, an NSF CAREER Award, and the US Presidential Early Career Award for Scientists and Engineers (PECASE). // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://cs.stanford.edu/~matei/ https://spark.apache.org/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Connect with Matei on LinkedIn: https://www.linkedin.com/in/mateizaharia/ Timestamps: [00:00] Matei's preferred coffee [01:45] Takeaways [05:50] Please subscribe to our newsletters, join our Slack, and subscribe to our podcast channels! [06:52] Getting to know Matei as a person [09:10] Spark [14:18] Open and freewheeling cross-pollination [16:35] Actual formation of Spark [20:05] Spark and MLFlow Similarities and Differences [24:24] Concepts in MLFlow [27:34] DJ Khalid of the ML world [30:58] Data Lakehouse [33:35] Stanford's unique culture of the Computer Science Department [36:06] Starting a company [39:30] Unique advice to grad students [41:51] Open source project [44:35] LLMs in the New Revolution [47:57] Type of company to start with [49:56] Emergence of Corporate Research Labs [53:50] LLMs size context [54:44] Companies to respect [57:28] Wrap up
MLOps Coffee Sessions #154 with Waleed Kadous, ML Scalability Challenges, co-hosted by Abi Aryan. // Abstract Dr. Waleed Kadous, Head of Engineering at Anyscale, discusses the challenges of scalability in machine learning and his company's efforts to solve them. The discussion covers the need for large-scale computing power, the importance of attention-based models, and the tension between big and small data. // Bio Dr. Waleed Kadous leads engineering at Anyscale, the company behind the open-source project Ray, the popular scalable AI platform. Prior to Anyscale, Waleed worked at Uber, where he led overall system architecture, evangelized machine learning, and led the Location and Maps teams. He previously worked at Google, where he founded the Android Location and Sensing team, responsible for the "blue dot" as well as ML algorithms underlying products like Google Fit. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: anyscale.com https://www.youtube.com/watch?v=hzW0AKKqew4https://www.anyscale.com/blog/WaleedKadous-why-im-joining-anyscale Ray Summit: https://raysummit.anyscale.com/ Anyscale careers: https://www.anyscale.com/careersLearning Ray O'Reilly book. It's free to anyone interested. https://www.anyscale.com/asset/book-learning-ray-oreilly --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Waleed on LinkedIn: https://www.linkedin.com/in/waleedkadous/ Timestamps: [00:00] Waleed's preferred coffee [00:38] Takeaways [07:37] Waleed's background [13:16] Nvidia investment with Rey [14:00] Deep Learning use cases [17:52] Infrastructure challenges [22:01] MLOps level of maturity [26:42] Scale overloading [29:21] Large Language Models [32:40] Balance between fine-tuning forces prompts engineering [35:51] Deep Learning movement [42:05] Open-source models have enough resources [44:11] Ray [47:59] Value add for any scale from Ray [48:55] "Big data is dead" reconciliation [52:43] Causality in Deep Learning [55:16] AI-assisted Apps [57:59] Ray Summit is coming up in September! [58:49] Anyscale is hiring! [59:25] Wrap up
This exclusive podcast episode covers the key findings from the LLM in-production survey that we conducted over the past month.
For all the data to explore yourself use this link https://docs.google.com/spreadsheets/d/13wdBwkX8vZrYKuvF4h2egPh0LYSn2GQSwUaLV4GUNaU/edit?usp=sharing
Sign up for our LLM in-production conference happening on April 13th (TODAY) here:
https://home.mlops.community/home/events/llms-in-production-conference-2023-04-13
MLOps Coffee Sessions #153 with Rodolfo Núñez, Multilingual Programming and a Project Structure to Enable It, co-hosted by Abi Aryan. // Abstract It's really easy to mix different programming languages inside the same project and use a project template that enables easy collaboration. It's not about what language is better, but rather what language solves the given section of your problem better for you. // Bio Rodo has been working in the "Data Space" for almost 7 years. He was a Senior Data Scientist at Entel (a Chilean telecommunications company) and is now a Senior Machine Learning Engineer at the same company, where I also lead three mini teams dedicated to internal cybersecurity; design/promote continuous training for the entire Analytics team and also the whole company; and ensure the improvement of programming practices and code cleanliness standards. Rodo is currently in charge of helping the team put models into production and define the tools that we will use for it. He specializes in R, but he's language/tool agnostic: you should use the tool that best solves your current problem. Rodo studied Mathematical Engineering and MSc in Applied Mathematics at the University of Chile in addition to General Engineering at the École Centrale Marseille. Rodo really likes to share knowledge (bi-directionally) in whatever he thinks he can contribute. Some things that Rodo like teaching are Data Science, Math, Latin Dances, and whatever he thinks he can give to people. Rodo's other interests are computer games (especially Vermintide and Darktide), board games, and dancing to Latin rhythms. Also, he streams some games and Data Science related topics on Twitch. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links https://www.twitch.tv/en_codershttps://www.youtube.com/@en_codershttps://www.twitch.tv/rodonunezhttps://github.com/rodo-nunezhttps://github.com/en-coders-cl --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Rodo on LinkedIn: https://www.linkedin.com/in/rodonunez/ Timestamps: [00:00] Rodo's preferred coffee [00:16] Project structure [00:34] Introduction to Rodolfo Núñez [01:20] Takeaways [04:34] Check out our Meetups, podcasts, newsletters, TikTok, and blog posts! [05:50] Why data scientists should know how to code and code properly [10:32] Becoming a team player [14:02] Cookie cutter project [17:50] Markdown and Quarter over Jupyter notebooks [23:18] Data scientists' templates [30:06] Significance of scripts [33:30] Monolith to Microservices [34:33] Reproducibility [36:37] Entire event processing scripts [40:44] In-House cataloging solution [42:08] Data flows [46:00] Bonus topics! [47:23] Elbow methodology [50:17] Idea behind cross sampling [50:51] Machine Learning and MLOps Security at Entel [58:04] Wrap up
Worlds are colliding! This week we join forces with the hosts of the Practical AI podcast to discuss all things machine learning operations. We talk about how the recent explosion of foundation models and generative models is influencing the world of MLOps, and we discuss related tooling, workflows, perceptions, etc.
--------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/
MLOps Coffee Sessions #152 with Keith Trnka, How A Manager Became a Believer in DevOps for Machine Learning. // Abstract Keith Trnka, a seasoned leader in the technology industry, set foot on the MLOps Podcast in a special episode where he shared insights into his experience leading data teams and machine learning teams, becoming a better software engineer, and overseeing a successful migration from a monolith to microservices in the healthcare sector without any downtime. Keith's background includes directing data science at 98.6, improving language models at swipe and nuance, and completing a Ph.D. thesis in language modeling for assistive technology. His work in these areas has contributed to the development of technology applications for healthcare, including telemedicine visits using natural language processing and machine learning. // Bio Keith has been in the industry for about 11 years. Most recently he was the Director of Data Science at 98point6 where we made telemedicine visits easier for doctors using natural language processing, machine learning, backend engineering, AWS, and frontend engineering. Prior to that, Keith improved the language models used in mobile phone keyboards at Swype and Nuance. And before that, He did his Ph.D. thesis in language modeling for assistive technology. Currently, Keith is traveling, mentoring, and doing a side project on machine translation. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Keith on LinkedIn: https://www.linkedin.com/in/keith-trnka/
MLOps Coffee Sessions #151 with Jean-Michel Daignan, ML in Production: A DS from Ubisoft Perspective, co-hosted by Abi Aryan. // Abstract As a data scientist himself, Jean-Michel has a unique perspective on the needs of data scientists when it comes to platform development. He talks about the non-invasive approach his team is taking to bring people onto the platform and their SDK, Merlin. The team is focused on tying machine learning products back to business use cases and the ROI they provide. Abby and Jean-Michel also discuss the use of generative AI and the importance of balancing delivering value and building things quickly. Jean-Michel's blog posts on the topic are recommended for further reading. // Bio The author of the blog "the-odd-dataguy.com" has been a data scientist for over 4.5 years at Ubisoft. Prior to joining the video game industry, Jean-Michel had a background in engineering from France and had previously worked in the energy sector. The blog focuses on topics related to data and machine learning, showcasing the author's expertise in the field. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Blog page: https://www.the-odd-dataguy.com/ Bringing Machine Learning to Production at Ubisoft (PydataMTL June22): https://www.the-odd-dataguy.com/2022/12/29/recap_pydata_mtl_june22/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Abi on LinkedIn: https://www.linkedin.com/in/goabiaryan/ Connect with Jean-Michel on LinkedIn: https://www.linkedin.com/in/jeanmicheldaignan/ Timestamps: [00:00] Jean-Michel's preferred beverage [00:19] Jean-Michel Daignan's background [00:28] Takeaways [04:30] Rate us and share the podcasts with your friends! [05:37] Jean-Michel's projects at Ubisoft [07:48] Jean-Michel's success as a Data Scientist [09:45] Ubisoft basics [10:40] Jean-Michel's success from the downfalls of being a data scientist [12:18] Building for data scientists' considerations [13:57] Differences in designing for data scientists in general [16:35] End twin pipelines and their functions [19:35] Major problems doing maintenance [20:53] Data quality ownership [22:33] Monitoring levels [24:25] Locomotive systems [26:14] Merlin [29:12] DS storage systems [31:09] Feature stores batch or streaming? [32:19] Bringing Machine Learning to Production at Ubisoft blog post [35:10] Features and recommendation systems [37:03] Playing games [38:21] Play data = play personalities [39:42] Deep learning in all the diffusion models or the foundation models [43:06] Servicing data scientists' needs [45:28] Ubisoft's data volume [48:00] Wrap up
LLM in Production Round Table with Demetrios Brinkmann, Diego Oppenheimer, David Hershey, Hannes Hapke, James Richards, and Rebecca Qian. // Abstract Using LLM in production. That's right. Hype or here to stay? The conversation answers some of the questions that have been asked by our community members like; performance & cost of production, the difference in architectures, Reliability issues, and a bunch of random tangents. We have some heavy hitters for this event! // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links LLM in Production survey: https://docs.google.com/forms/d/e/1FAIpQLSerEryK4xHEZTq0hSu-sVmBHilOzaT71BfCQgXe_uIRgIah-g/viewform Virtual LLMs in Production Conference registration: https://home.mlops.community/public/events/llms-in-production-conference-2023-04-13 Chinchilla papers: https://paperswithcode.com/method/chinchilla, https://arxiv.org/abs/2203.15556 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Diego on LinkedIn: https://www.linkedin.com/in/diego/ Connect with David on LinkedIn: https://www.linkedin.com/in/david-hershey-458ab081/ Connect with Hannes on LinkedIn: https://www.linkedin.com/in/hanneshapke/ Connect with James on LinkedIn: https://www.linkedin.com/in/james-richards-4baa73a7/ Connect with Rebecca on LinkedIn: https://www.linkedin.com/in/rebeccaqian/ Timestamps: [00:00] Round table success to Virtual LLM in Production Conference on April 13th! [00:18] Register for the Virtual LLM in Production Conference now! [00:44] LLM in Production survey [01:40] Lightning round of introduction of speakers [04:34] Large Language Models definition [09:17] What do we consider large? [10:35] Thought process in use cases production [14:30] LLM open source huge movements [16:50] Problems qualifications [19:25] Production use cases frameworks directions [25:25] Open-source language models tokenizer [26:25] Language models democratization [29:25] Three categories for LLMs in Production [31:22] Latency at 2 levels [33:27] Defining production [34:57] Hitting the latency problems [38:20] Fundamental latency barrier [40:39] Latency use case requirement [44:25] Costs and the use cases [48:12] Product management involvement in costing [49:38] LLMs Hallucination definition [52:05] Building deterministic systems trust [55:21] Wrap up
MLOps Coffee Sessions #150 with Saahil Jain, The Future of Search in the Era of Large Language Models, co-hosted by David Aponte.
// Abstract Saahil shares insights into the You.com search engine approach, which includes a focus on a user-friendly interface, third-party apps, and the combination of natural language processing and traditional information retrieval techniques. Saahil highlights the importance of product thinking and the trade-offs between relevance, throughput, and latency when working with large language models.
Saahil also discusses the intersection of traditional information retrieval and generative models and the trade-offs in the type of outputs they produce. He suggests occupying users' attention during long wait times and the importance of considering how users engage with websites beyond just performance.
// Bio Saahil Jain is an engineer at You.com. At You.com, Saahil builds searching and ranking systems. Previously, Saahil was a graduate researcher in the Stanford Machine Learning Group under Professor Andrew Ng, where he researched topics related to deep learning and natural language processing (NLP) in resource-constrained domains like healthcare. His research work has been published in machine learning conferences such as EMNLP, NeurIPS Datasets & Benchmarks, and ACM-CHIL among others. He has publicly released various machine learning models, methods, and datasets, which have been used by researchers in both academic institutions and hospitals across the world, as part of an open-source movement to democratize AI research in medicine. Prior to Stanford, Saahil worked as a product manager at Microsoft on Office 365. He received his B.S. and M.S. in Computer Science at Columbia University and Stanford University respectively. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Website: http://saahiljain.me/ --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/ Connect with Saahil on LinkedIn: https://www.linkedin.com/in/saahiljain/ Timestamps [00:00] Saahil's preferred coffee [04:32] Saahil Jain's background [04:44] Takeaways [07:49] Search Landscape [12:57] Use cases exploration [14:51] Differentiating what to give to users [17:19] Search key challenges [20:05] Search objective relevance [23:22] MLOps Search and Recommender Systems [26:54] Addressing Latency Issues [29:41] Throughput presenting results [32:20] Compute challenges [34:24] Working at a small start-up [36:10] Citations critics [39:17] Use cases to build [40:40] Integrating to Leveraging You.com [42:26] Open AI [46:13] Interfacing with bugs [49:16] Staying focused [52:05] Retrieval augmented models [52:32] Closing thoughts [53:47] Wrap up
MLOps Coffee Sessions #149 with Jason McCampbell, The Challenges of Deploying (many!) ML Models, co-hosted by Abi Aryan and sponsored by Wallaroo.
// Abstract
In order to scale the number of models a team can manage, we need to automate the most common 90% of deployments to allow ops folks to focus on the challenging 10% and automate the monitoring of running models to reduce the per-model effort for data scientists. The challenging 10% of deployments will often be "edge" cases, whether CDN-style cloud-edge, local servers, or running on connected devices.
// Bio
Jason McCampbell is the Director of Architecture at Wallaroo.ai and has over 20 years of experience designing and building high-performance and distributed systems. From semiconductor design to simulation, a common thread is that the tools have to be fast, use resources efficiently, and "just work" as critical business applications.
At Wallaroo, Jason is focused on solving the challenges of deploying AI models at scale, both in the data center and at "the edge". He has a degree in computer engineering as well as an MBA and is an alum of multiple early-stage ventures. Living in Austin, Jason enjoys spending time with his wife and two kids and cycling through the Hill Country.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://wallaroo.ai
MLOps at the Edge Slack channel: https://mlops-community.slack.com/archives/C02G1BHMJRL
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Jason on LinkedIn: https://www.linkedin.com/in/jasonmccampbell/
Timestamps:
[00:00] Jason's preferred coffee
[01:22] Takeaways
[06:06] MLOps at the Edge Slack channel
[06:36] Shoutout to Wallaroo!
[07:34] Jason's background
[09:54] Combining Edge and ML
[11:03] Defining Edge Computing
[13:21] Data transport restrictions
[15:02] Protecting IP in Edge Computing
[17:48] Decentralized Teams and Knowledge Sharing
[20:45] Real-time Data Analysis for Improved Store Security and Efficiency
[24:49] How to Ensure Statistical Integrity in Federated Networks
[26:50] Architecting ML at the Edge
[30:44] Machine Learning models adversarial attacks
[33:25] Pros and cons of baking models into containers
[34:52] Jason's work at Wallaroo
[38:22] Navigating the Market Edge
[40:49] Last challenges to overcome
[44:15] Data Science Use Cases in Logistics
[46:27] Vector trade-offs
[49:34] AI at the Edge challenges
[50:40] Finding the Sweet Spot
[53:46] Driving revenue
[55:33] Wrap up
MLOps Coffee Sessions #148 with Karl Fezer, Intelligence & MLOps co-hosted by Abi Aryan.
// Abstract
This conversation explores various topics including biases, defining intelligence, and the future of large language models and MLOps. Karl discusses his paper on defining intelligence and how it relates to the increasing interest in Artificial Intelligence. Karl shares his thoughts on the overlap between foundational models and MLOps, emphasizing the importance of making high-impact tasks more efficient and easier. The conversation touched on philosophical tangents but ultimately circled back to practical applications of these concepts.
// Bio
Karl graduated in 2014 from the University of Georgia with a Masters in Science in Artificial Intelligence. Since then, he has continued to stay on top of the latest iterations of Machine Learning and loves trying new open-source frameworks.
For the last 6 years, he has been purely focused on AI Developer Relations. First at Mycroft, the voice assistant startup, then Intel, Arm, briefly at Wallaroo.ai, and now at Lockheed Martin.
He currently lives in Seattle and spends his free time writing, reading, sailing, and camping.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Intelligence Primer paper by Karl Fezer and Andrew Sloss: https://arxiv.org/abs/2008.07324
Computing Machinery and Intelligence paper by Alan Turing: https://redirect.cs.umbc.edu/courses/471/papers/turing.pdf
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Karl on LinkedIn: https://www.linkedin.com/in/karlfezer/
MLOps Coffee Sessions #147 with Alex DeBrie, Something About Databases co-hosted by Abi Aryan.
// Abstract
For databases, it feels like we're in the middle of a big shift. The first 10-15 years of the cloud were mostly about using the same core infrastructure patterns but in the cloud (SQL Server, MySQL, Postgres, Redis, Elasticsearch).
In the last few years, we're finally seeing data infrastructure that is truly built for the cloud. Elastic, scalable, resilient, managed, etc. Early examples were Snowflake + DynamoDB. The most recent ones are all the 'NewSQL' contenders (Cockroach, Yugabyte, Spanner) or the 'serverless' ones (Neon, Planetscale). Also seeing improvements in caching, search, etc. Exciting times!
// Bio
Alex is an AWS Data Hero and self-employed AWS consultant and trainer. He is the author of The DynamoDB Book, a comprehensive guide to data modeling with DynamoDB. Previously, he worked for Stedi and for Serverless, Inc., creators of the Serverless Framework. He loves being involved in the AWS & serverless community, and he lives in Omaha, NE with his family.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related
Links Key Takeaways from the DynamoDB Paper: https://www.alexdebrie.com/posts/dynamodb-paper/
Understanding Eventual Consistency in DynamoDB: https://www.alexdebrie.com/posts/dynamodb-eventual-consistency/
Two Scoops of Django 1.11: Best Practices for the Django Web Framework: https://www.amazon.com/Two-Scoops-Django-1-11-Practices/dp/0692915729CAP or no CAP?
Understanding when the CAP theorem applies and what it means: https://www.alexdebrie.com/posts/when-does-cap-theorem-apply/
Stop fighting your database/ The DynamoDB book: https://dynamodbbook.com/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Alex on LinkedIn: https://www.linkedin.com/in/alex-debrie/
Timestamps:
[00:00] Alex's preferred coffee
[00:27] Introduction to Alex DeBrie and DynamoDB
[01:05] Takeaways
[03:47] Please write down your reviews and you might have a book of Alex!
[04:57] Alex's journey from being an Attorney to becoming a Data Engineer
[07:31] Why engineering?
[10:07] Serverless Data
[12:54] Before Airflow
[15:41] Batch vs streaming
[17:03] Difficulties in Batch and Streaming side
[19:45] Modern data infrastructure databases
[24:37] Cloud Ecosystem Maturity
[27:59] New generation type of Snowflake
[29:47] Comparing databases
[30:58] What's next on connectors from 2 perspectives?
[34:25] Management services at the MLOps level
[36:38] DynamoDB
[39:32] Why do you like DynamoDB?
[41:00] Data used in DynamoDB and size limits
[43:46] Comparison of tradeoffs between DynamoDB and Redis
[45:52] Preferred opinionated databases
[48:43] CAP or no CAP? Understanding when the CAP theorem applies and what it means
[52:10] The DynamoDB book
[56:17] Chapter you want to expand on the book
[57:43] Next book to write
[59:25] ChatGPT iterations
[1:01:59] Data modeling book wished to be written
[1:03:27] Wrap up
MLOps Coffee Sessions #146 with Shalabh Chaudri, The Ops in MLOps - Process and People co-hosted by Abi Aryan.
// Abstract
Shalabh talks through their newfound appreciation for the MLOps perspective from a customer success standpoint. Shalabh's emphasis on setting realistic expectations and ensuring the delivery of promised value adds is particularly valuable.
Generally, this episode provides a unique and insightful perspective on MLOps from the lens of customer success.
// Bio
Shalabh has worked in the MLOps domain since 2020 at Algorithmia and Union AI. His experience spans startups and small and large public companies. He has 10+ years of experience in the design, delivery, adoption, and business value realization of B2B infrastructure and platform solutions.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://www.union.ai/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Shalabh on LinkedIn: https://www.linkedin.com/in/shalabhchaudhri/
Timestamps:
[00:00] Shalabh's preferred coffee
[01:18] Takeaways
[02:57] Huge shout out to Union AI!
[03:46] Reviews
[05:26] Shalab's journey
[07:00] The people and process of MLOps
[10:25] Accuracy measures and Multiple Stakeholders
[13:01] UnionAI's success where others fall short
[14:45] Legacy equipment
[17:06] Legacy tools versus open source
[19:27] Cataloging solution
[22:51] Stakeholders and maturity levels
[24:26] People and Process in MLOps
[29:00] Collaboration for Machine Learning
[31:08] Overcoming challenges
[34:17] AI and leadership decision-making
[35:33] Legacy Companies and AI
[39:39] Common pitfalls
[42:24] Neglecting ROI
[46:25] Speaking to each level
[49:50] Being realistic
[51:29] Becoming a champion
[53:08] Transitioning to machine learning
[55:25] Customer's Skill and Success needed in ML
[57:46] Different sizes of companies
MLOps Coffee Sessions #145 with Sahil Khanna, Griffin, ML platform at Instacart co-hosted by Mike Del Balso.
// Abstract
The conversation revolves around the journey of Instacart in implementing machine learning, starting from batch processing to real-time processing. The speaker highlights the importance of real-time processing for businesses and the relevance of Instacart's journey to other machine learning teams.
Sahil emphasizes the soft factors, such as staying customer-focused and the right approach, that contributed to the success of Instacart's machine learning implementation. We also recommend two blog posts by Sahil about Instacart's journey.
// Bio
Sahil is currently a machine learning engineer at Instacart, where they are building a centralized platform for the training, deployment, and management of diverse ML applications. Before Instacart, Sahil developed ML training and inference platforms at Etsy.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mike on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/
Connect with Sahil on LinkedIn: www.linkedin.com/in/sahil-khanna-umd
MLOps Coffee Sessions #144 with Matthew Dombrowski, Non-traditional Career Paths in MLOps co-hosted by Mihail Eric.
// Abstract
Let's explore the different aspects of ML and data roles and the variety of responsibilities each role entails! This conversation emphasizes the need for understanding the unique insights each role provides and the similarities in responsibilities and soft skills that are required across different roles.
This episode also highlights the significance of stakeholder alignment in the context of working in big companies and the importance of navigating these complexities for a successful career in ML.
// Bio
Matt has performed a number of MLOps positions including Solutions Consultant, Solutions Architect, and Product Manager from startups to large organizations. In his current role, Matt builds tools to help social media influencers discover unique and exciting Amazon products to recommend to their audiences.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Matt on LinkedIn: https://www.linkedin.com/in/matthewdombrowski/
Timestamps:
[00:00] Matt's preferred coffee
[00:28] Mihail's new creation
[05:09] Introduction to Matthew Dombrowski
[06:02] Takeaways
[08:30] Pizza and coffee nerds
[10:54] Data careers
[13:35] Matt's progression through the ml sphere
[20:10] Dealing with machine learning
[23:20] Transition from deep technical implementer to PM role
[27:42] Data is a product
[29:30] From start-ups to big companies
[32:41] Ambiguity of ML
[36:17] Matt's daily routine
[40:23] Social media influencers
[42:07] Developer advocate
[44:00] Stakeholder alignment
[49:41] Non-traditional career paths military influence
[54:11] Good ways to recommend people to get into ML
[57:56] MLOps Meetups all over the world
[59:00] Wrap up
MLOps Coffee Sessions #143 with Jill Chase & Manmeet Gujral, Investing in the Next Generation of AI & ML.
// Abstract
Investors are currently focusing on developer tooling and the foundational AI model movement, as they have seen explosive growth in this area. This podcast explores the impact of foundational models on investment thesis and the future of machine learning operations.
The discussion also touches on the idea of generative AI and large language models, and their potential impact on MLOps in the next 10 years. Jill and Manmeet from Capital G share their insights on this topic.
// Bio
Jill Chase
Jill is an investor at CapitalG where she focuses on enterprise software, with an emphasis on data infrastructure and AI/ML.
Prior to joining CapitalG, Jill worked in senior startup operating roles, both as the CEO of a private equity-backed business and as the founder of a Y Combinator-backed startup.
Jill graduated magna cum laude from Williams College with a dual degree in Economics and Psychology and was captain of the women’s basketball team. She came out to the West Coast to earn an MBA from the Stanford Graduate School of Business, but she was born and raised in Boston where she had the opportunity to cheer on the most impressive era of professional sports a city has ever experienced (Go Patriots).
She lives in the Bay Area with her husband where they spend weekends doing as many outside activities as possible, such as pickleball, tennis, hiking, and running.
Manmeet Gujral
Manmeet is a member of the CapitalG investment team where he focuses on enterprise software, AI & ML, open source, and product-led-growth companies. Prior to joining CapitalG in 2021, Manmeet worked in product marketing and operations at Tecton. Before that, he worked as a consultant at Bain & Company in San Francisco where he specialized in the go-to-market strategy for technology companies and private equity investment diligence. Manmeet is originally from Albany, New York, and graduated from Dartmouth College with a dual degree in Computer Science and Economics. Manmeet is highly opinionated about pizza, an avid New York sports fan, and always willing to share his latest house or hip-hop playlists.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Jill on LinkedIn: https://www.linkedin.com/in/jill-greenberg-chase-53747538/
Connect with Manmeet on LinkedIn: https://www.linkedin.com/in/manmeet-gujral/
Timestamps:
[00:00] Manmeet and Jill's preferred coffee
[00:25] Takeaways
[01:31] CapitalG, Jill and Manmeet's Background
[05:12] Sideswiping MLOps by Foundational Models
[08:50] MLOps space and the market revenue
[14:50] Foundational models B to C style
[20:37] Foundational models taking over
[23:00] Uninnovative sentiments
[27:50] 2 prototypes of companies
[31:51] Finding product market fit
[36:20] MLOps market growth changes
[40:30] Monster valuations
[41:43] The ones that got away
[44:07] Wrap up
MLOps Coffee Sessions #142 with Murtuza Shergadwala, Approaches to Fairness and XAI co-hosted by Abi Aryan. This episode is sponsored by Fiddler AI.
// Abstract
The field of Explainable Artificial Intelligence (XAI) is continuously evolving, with an increasing focus on providing model-centric explanations in a human-centric manner. However, better frameworks and training for users are needed to fully utilize the potential of XAI tools.
Additionally, there is a discrepancy in the approach to fairness in XAI, with the industry approaching it from a regulatory standpoint, while academia is engaging in more discussion and research on the topic.
// Bio
Dr. Murtuza Shergadwala is a data scientist at Fiddler AI. His background is in human-machine interaction and design decision-making. He received his Ph.D. from Purdue University in Mechanical Engineering. Prior to Fiddler, he was a postdoc at the Games User Interaction and Intelligence Lab at UC Santa Cruz where he focused on using Bayesian approaches for modeling cognition and investigating the theory of mind. He’s super passionate about fairness in AI for underrepresented communities.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://murtuzashergadwala.wixsite.com/murtuza
https://www.fiddler.ai/blog/detecting-intersectional-unfairness-in-ai-part-1
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Murtuza on LinkedIn: https://www.linkedin.com/in/murtuza-shergadwala/
Timestamps:
[00:00] Moto's preferred coffee
[00:35] Introduction to Murtuza Shergadwala
[01:06] Takeaways
[04:30] Huge shout out to Fiddler AI for sponsoring this episode!
[05:00] Don't forget to like, comment, and subscribe. Give us a rating
[06:10] Moto's background and transition to Human-centric AI
[10:52] Decision-making behaviors of engineering designers in design contests
[15:10] Gaining insights from data decisions
[18:00] Defining latent variables
[20:32] Designer's perspective on building systems
[23:14] XAI as a movement
[27:47] Selling regulations and bridging the gap
[32:18] Data integrity towards detecting outliers alerting and data drifts
[34:32] Dealing with alerts and alert fatigue
[37:31] Approaches and their limitations
[39:10] Alert-level systems
[42:19] Alerts putting into practice
[45:30] Creative alerts
[47:02] One solution fits all?
[50:08] Wrap up
MLOps Coffee Sessions #141 with Stephen Bailey, Airflow Sucks for MLOps co-hosted by Joe Reis.
// Abstract
Stephen discusses his experience working with data platforms, particularly the challenges of training and sharing knowledge among different stakeholders. This talk highlights the importance of having clear priorities and a sense of practicality and mentions the use of modular job design and data classification to make it easier for end users to understand which data to use.
Stephen also mentions the importance of being able to move quickly and not getting bogged down in the quest for perfection. We recommend Stephen's blog post "Airflow's Problem" for further reading.
// Bio
Stephen has worked as a data scientist, analyst, manager, and engineer, and loves all the domains equally. He currently works at Whatnot, a collectibles marketplace that focuses on live shopping, and has previously worked in privacy tech at Immuta. He has his Ph.D. from Vanderbilt University in educational cognitive neuroscience, but it has yet to help him understand why his three children are so crazy.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Airflow's Problem blog post: https://stkbailey.substack.com/p/airflows-problem
Airflow's Problem and the reception it got on Hacker News: https://news.ycombinator.com/item?id=32317558
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Joe on LinkedIn: https://www.linkedin.com/in/josephreis/
Connect with Stephen on LinkedIn: https://www.linkedin.com/in/stkbailey/
Timestamps:
[00:00] Stephen's preferred coffee
[00:19] Introduction to co-host Joe Reis
[01:40] Takeaways
[06:29] Subscribe to our newsletters!
[06:55] Shout out to our sponsor, Wallaroo!
[08:05] Whatnot
[10:47] Stephen's side hustle
[14:35] Stephen's work breakdown at Whatnot
[18:03] Fundamental tensions in the data world
[21:27] Initial questions to answer that you were on the right path
[24:06] Recommender systems
[28:15] Coordinating with ML teams
[29:43] Daxter
[31:38] Too advanced, more challenging
[34:37] Orchestration layer
[36:14] Decision criteria
[39:23] Human design aspect of Daxter
[40:53] Orchestration layer centralization and sharing knowledge with stakeholders
[46:18] Airflow's Problem and the reception it got on Hacker News
[51:00] Wrap up
MLOps Coffee Sessions #140 with Sakib Dadi, The Evolution of ML Infrastructure sponsored by Wallaroo.
// Abstract
The toolkit and infrastructure empowering machine learning practitioners are advancing as ML adoption accelerates. We'll go through the current landscape of ML tooling, startups, and new projects from an investor's perspective.
// Bio
Sakib is a vice president in the San Francisco office where he primarily focuses on early-stage investments in developer platforms, data infrastructure, and machine learning. He has been involved with Bessemer’s investments in Prefect, Coiled, Arcion, Periscope Data (acquired by Sisense), Okera, npm (acquired by GitHub), and LaunchDarkly. Before joining Bessemer, Sakib worked in product at Viagogo, an international marketplace for buying and selling tickets for live events.
Sakib also worked in the technology investment banking group at Morgan Stanley and as an engineer at Innova Dynamics (acquired by TPK), a startup manufacturing flexible touchscreen displays.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Sakib on LinkedIn: https://www.linkedin.com/in/sakib-dadi-77938937/
Timestamps:
[00:00] Sakib's preferred coffee
[00:13] Introduction to Sakib Dadi
[01:33] Sakib's background
[02:40] Shout out to this episode's Sponsor, Wallaroo!
[04:17] ML investing
[05:57] Investing regrets
[08:06] Transformers are today what would be tomorrow?
[09:18] Company that you wish existed now
[10:23] Current thoughts on the MLOps market
[12:32] MLOps transition to Generative AI
[15:52] Mind maps
[17:03] Jasper
[22:14] Intersection
[24:10] Differences in models in-house
[26:08] Orchestration space
[28:23] Nuances of Monitoring
[29:20] Demetrios' theory on Monitoring
[31:48] Non-funded Monitoring Companies
[34:29] Investment risks
[36:55] Orchestration markets
[39:38] MLOps market at a plateau
[42:14] Vertical problems, vertical solutions
[45:45] Sakib starting a company
[50:10] Structuring deals
[51:50] Infrastructure tools
[53:26] Firing a founder
[53:49] Parting ways with a founder
[57:07] Wrap up
MLOps Coffee Sessions #139 with Alex Ratner, Putting Foundation Models to Use for the Enterprise co-hosted by Abi Aryan sponsored by Snorkel AI.
// Abstract
Foundation models are rightfully being compared to other game-changing industrial advances like steam engines or electric motors. They’re core to the transition of AI from a bespoke, less predictable science to an industrialized, democratized practice. Before they can achieve this impact, however, we need to bridge the cost, quality, and control gaps.
Snorkel Flow Foundation Model Suite is the fastest way for AI/ML teams to put foundation models to use. For some projects, this means fine-tuning a foundation model for production dramatically faster by creating programmatically labeling training data. For others, the optimal solution will be using Snorkel Flow’s distill, combine, and correct approach to extract the most relevant knowledge from foundation models and encode that value into the right-sized models for your use case.
AI/ML teams can determine which Foundation Model Suite capabilities to use (and in what combination) to optimize for cost, quality, and control using Snorkel Flow’s integrated workflow for programmatic labeling, model training, and rapid-guided iteration.
// Bio
Alex Ratner is the Co-founder and CEO of Snorkel AI and an Assistant Professor of Computer Science at the University of Washington.
Prior to Snorkel AI and UW, he completed his Ph.D. in CS advised by Christopher Ré at Stanford, where he started and led the Snorkel open-source project, and where his research focused on applying data management and statistical learning techniques to emerging machine learning workflows such as creating and managing training data and applying this to real-world problems in medicine, knowledge base construction, and more. Previously, he earned his A.B. in Physics from Harvard University.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: www.snorkel.ai
Huge “foundation models” are turbo-charging AI progress: https://www.economist.com/interactive/briefing/2022/06/11/huge-foundation-models-are-turbo-charging-ai-progress
Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming: https://arxiv.org/abs/2203.01382
The Principles of Data-Centric AI Development: https://snorkel.ai/principles-of-data-centric-ai-development/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Alex on LinkedIn: https://www.linkedin.com/in/alexander-ratner-038ba239/
Timestamps: [00:00] Alex's preferred coffee [01:20] Introduction to Alex Ratner [02:34] Takeaways [04:04] Huge shoutout to our Sponsor, Snorkel AI! [04:39] Comment, rate us, and share our podcasts with your friends! [04:50] Transfer Learning / Active Learning [11:30] Labeling Heuristics paper on Nemo [18:14] Geocentric AI [21:48] Enterprise use cases on Foundational Models [32:45] Foundational Models into the different Google products [38:36] Progress in Foundational Models [43:55] AutoML Models Baseline Accuracy [44:40] Hosting Infrastructure Snorkel Float vs GCP [46:53] Chris Re's venture capital firm / incubator / machine [51:00] Wrap
MLOps Coffee Sessions #138 with Dattaraj Rao, Explainability in the MLOps Cycle co-hosted by Vishnu Rachakonda.
// Abstract
When it comes to Dattaraj's interest, you'll hear about his top 3 areas in Machine Learning. What he sees as up and coming, what he's investing his company's time into and where he invests his own time.
Learn more about rule-based systems, deploying rule-based systems , and how to incorporate systems into more systems. there is no difference between ML systems and deploying models. It's just that this machine learning model is much smarter than traditional rule based models.
// Bio
Dattaraj Jagdish Rao is the author of the book “Keras to Kubernetes: The Journey of a Machine Learning Model to Production”. Dattaraj leads the AI Research Lab at Persistent and is responsible for driving thought leadership in AI/ML across the company. He leads a team that explores state-of-the-art algorithms in Knowledge Graphs, NLU, Responsible AI, MLOps and demonstrates applicability in Healthcare, Banking, and Industrial domains. Earlier, he worked at General Electric (GE) for 19 years building Industrial IoT solutions for Predictive Maintenance, Digital Twins, and Machine Vision.
Dattaraj held several Technology Leadership roles at Global Research, GE Power, and Transportation (now part of Wabtec). He led the Innovation team out of Bangalore that incubated video track inspection from an idea into a commercial Product. Dattaraj has 11 patents in Machine Learning and Computer Vision areas.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Keras to Kubernetes: The Journey of a Machine Learning Model to Production book:
https://www.amazon.com/Keras-Kubernetes-Journey-Learning-Production/dp/1119564832
Responsible Data Science Research | Talk @ VLDB 2022| Dattaraj Rao
https://www.youtube.com/watch?v=5_19KvSiy8s
Operationalizing AI/ML: Journey of an ML Model to Production | Masterclass by Dattaraj Rao
https://www.youtube.com/watch?v=Zk3RiiG07Us
Dattaraj Rao presenting workshop on MLOps at VISUM 2021
https://www.youtube.com/watch?v=wonUvbMDTUA
Machine Learning Design Patterns book:
https://www.oreilly.com/library/view/machine-learning-design/9781098115777/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Dattaraj on LinkedIn: https://www.linkedin.com/in/dattarajrao/
MLOps Coffee Sessions #137 with Niklas Kühl, Machine Learning Operations — What is it and Why Do We Need It? co-hosted by Abi Aryan.
// Abstract
The final goal of all industrial machine learning (ML) projects is to develop ML products and rapidly bring them into production.
However, it is highly challenging to automate and operationalize ML products and thus many ML endeavors fail to deliver on their expectations. The paradigm of Machine Learning Operations (MLOps) addresses this issue.
// Bio
NIKLAS KÜHL studied Industrial Engineering & Management at the Karlsruhe Institute of Technology (KIT) (Bachelor and Master). During his studies, he gained practical experience in IT by working at Porsche in both national and international roles. Niklas has been working on machine learning (ML) and artificial intelligence (AI) in different domains since 2014. In 2017, he gained his PhD (summa cum laude) in Information Systems with a focus on applied machine learning from KIT. In 2020, he joined IBM.
As of today, Niklas engages in two complementary roles: He is head of the Applied AI in Services Lab at the Karlsruhe Institute of Technology (KIT), and, furthermore, he works as a Managing Consultant for Data Science at IBM. In his academic and practical projects, he is working on conceptualizing, designing, and implementing AI in Systems with a focus on robust and fair AI as well as the effective collaboration between users and intelligent agents. Currently, he and his team are actively working on different ML & AI solutions within industrial services, sales forecasting, production lines or even creativity. Niklas is internationally collaborating with multiple institutions like the University of Texas and the MIT-IBM Watson AI Lab.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: niklas.xyz
MLOps Newsletters: https://airtable.com/shrx9X19pGTWa7U3Y
Machine Learning Operations (MLOps): Overview, Definition, and Architecture paper: https://arxiv.org/abs/2205.02302
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Niklas on LinkedIn: https://www.linkedin.com/in/niklaskuehl/
Timestamps:
[00:00] Niklas' preferred coffee
[00:43] Introduction to Niklas Kühl
[01:16] Takeaways
[02:05] Subscribe to our newsletters and give us a rating here!
[02:54] Niklas background
[05:09] Scraping twitter data
[06:58] EV's conclusions
[08:24] NLP usage on Twitter
[10:26] Consumer behavior production
[12:03] Management and Machine Learning Systems Communication
[14:00] Current hype around Machine Learning
[15:10] Budgeting ML Productions
[18:15] Machine Learning Operations (MLOps): Overview, Definition, and Architecture paper
[22:56] Niklas' MLOps definiton
[25:55] Navigating the idea of MLOps
[30:34] Return of Investment endeavor
[33:58] Full stack data scientist
[37:39] Defining success for different kinds of data science projects [41:06] Fun fact about Niklas [44:35] Other things Niklas do [47:02] The world is your oyster [50:57] Niklas' day to day life [52:48] One lecture Niklas can drop in [53:57] Foundational models [58:20] Wrap up
MLOps Coffee Sessions #136 with Andrew Dye, Systems Engineer Navigating the World of ML co-hosted by David Aponte.
// Abstract
We don't hear that much about working at a very low level on this podcast but they are still very valid. Andrew is able to give us his take on why and what you need to keep in mind when you are working at these low levels and why it is very important when you are a Machine Learning Engineer and how the two can play together nicely.
Most MLOps teams are formed using existing people and exitsing engineers. More often than not you have to blend these various disciplines and it works well when there's a common goal.
// Bio
Andrew is a software engineer at Union and contributor to Flyte, a production grade data and ML orchestration platform. Prior to that he was a tech lead for ML Infrastructure at Meta, where he focused on ML training reliability.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Andrew on LinkedIn: https://www.linkedin.com/in/andrewwdye
Timestamps:
[00:00] Andrew's preferred coffee
[03:30] Introduction to Andrew Dye
[03:33] Takeaways
[07:32] Huge shoutout to our sponsors UnionML and UnionAI!
[07:48] Andrew's background
[10:08] Andrew's learning curve
[11:10] Bridging the gap between firmware space and MLOps
[12:18] In connection with Pytorch team
[12:54] Things that should have learned sooner
[14:54] Type of scale Andrew works on
[17:42] Distributed training at Meta
[19:55] Managing the huge search space
[22:18] Execution patterns programs
[23:20] Non-ML engineers dealing with ML engineers having the same skill set
[26:44] Pace rapid change adoptation
[29:18] Consensus challenges
[32:26] Abstractions making sense now
[34:53] Comparing to others
[39:21] General principles in UnionAI tooling
[41:54] Seeing the future
[43:54] Inter-task checkpointing
[44:52] Combining functionality with use cases
[46:17] Wrap up
MLOps Coffee Sessions #135 with Sasha Ovsankin and Rupesh Gupta, Real-time Machine Learning: Features and Inference co-hosted by Skylar Payne.
// Abstract
Moving from batch/offline Machine Learning to more interactive "near" real-time requires knowledge, team, planning, and effort. We discuss what it means to do real-time inference and near-real-time features when to do this move, what tools to use, and what steps to take.
// Bio
Sasha Ovsankin Sasha is currently a Tech Lead of Machine Learning Model Serving infrastructure at LinkedIn, worked also on Feathr Feature Store, Real-Time Feature pipelines, designed metric platforms at LinkedIn and Uber, and was co-founder in two startups. Sasha is passionate about AI, Software Craftsmanship, improvisational music, and many more things.
Rupesh Gupta
Rupesh is a Sr. Staff Engineer in the AI team at LinkedIn. He has 10 years of experience in search and recommender systems.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Sasha on LinkedIn: https://www.linkedin.com/in/sashao/
Connect with Rupesh on LinkedIn: https://www.linkedin.com/in/guptarupesh
Timestamps:
[00:00] Sasha's and Rupesh's preferred coffee
[01:30] Takeaways
[07:23] Changes in LinkedIn
[09:21] "Real-time" Machine Learning in LibnkedIn
[13:08] Value of Feedback
[14:24] Technical details behind getting the most recent information integrated into the models
[16:53] Embedding Vector Search action occurrence
[18:33] Meaning of "Real-time" Features and Inference
[20:23] Are "Real-time" Features always worth that effort and always helpful?
[23:22] Importance of model application
[25:26] Challenges in "Real-time" Features
[30:40] System design review on Pinterest
[36:13] Successes of real-time features
[38:31] Learnings to share
[45:52] Branching for Machine Learning
[48:44] Not so talked about discussion of "Real-time"
[51:09] Wrap up
MLOps Coffee Sessions #134 with Jeremy Thomas Jordan, Building Threat Detection Systems: An MLE's Perspective co-hosted by Vishnu Rachakonda.
// Abstract
There is a clear pattern that we have been seeing with some of these greats in MLOps. So many use writing as a forcing function to learn about where they have holes in their understanding of something.
If you are not writing, this episode is important as to why writing is important for your own development. Jeremy goes into writing in depth as to how beneficial it is for him to write and for him to see that he doesn't understand something if he cannot re-articulate it in writing.
// Bio
Jeremy is a machine learning engineer currently working at Duo Security where he focuses on building ML infrastructure to operate threat detection systems at scale. He previously worked at Proofpoint, where he built models for phishing and malware detection.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links Website:
https://www.jeremyjordan.me/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Visnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Jeremy on Twitter: https://twitter.com/jeremyjordan
MLOps Coffee Sessions #133 {Podcast BTS} with Chip Huyen, Real-time Machine Learning with Chip Huyen co-hosted by Vishnu Rachakonda.
// Abstract
Forcing functions and how you can supercharge your learning by putting yourself into a situation where you know you either have a responsibility to others to learn or accountability on you so you have to learn.
It's not that hard when you think about streaming machine learning. It's not that big of a mental barrier to cross. It is simple in theory but maybe it's more complicated in practice and that's exactly where Chip's perspective is.
// Bio
Chip Huyen is a co-founder of Claypot AI, a platform for real-time machine learning. Previously, she was with Snorkel AI and NVIDIA. She teaches CS 329S: Machine Learning Systems Design at Stanford. She’s the author of the book Designing Machine Learning Systems, an Amazon bestseller in AI.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Landing page: https://claypot.ai
Designing Machine Learning Systems book:
https://www.amazon.com/Designing-Machine-Learning-Systems-Production-Ready/dp/1098107969
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Chip on LinkedIn: https://www.linkedin.com/in/chiphuyen/
MLOps Coffee Sessions #132 {Podcast BTS} with Ian Schweer, What is Data / ML Like on League? co-hosted by Skylar Payne.
// Abstract
If you're not an avid gamer yourself, you have never really seen how ML might be used in the gaming space. It's so interesting to see the things that are different like full stories of players' games from start to finish.
// Bio
On the surface, Ian is an excellent developer who gets things done. Underneath, he is much more. Ian is a reliable and trustworthy teammate who demonstrates an exceptional ownership mentality.
Here's a fair share of Ian's job history:
2014 - UCI (With Skylar!)
2015 - Adobe Primetime (SWE)
2017 - Adobe Product and Customer Analytics (SWE)
2019 - Doordash Data Infra (SWE) Current - Riot Games on League
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links Landing page:
https://www.riotgames.com/en
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Ian on LinkedIn: https://www.linkedin.com/in/ianschweer/
Timestamps:
[00:00] Ian's preferred coffee
[02:10] Takeaways
[05:14] Please hit the like button and leave us a review. Please subscribe also!
[05:45] Engineering Community Mental Health Awareness
[07:33] Coping mechanism
[09:29] Increase in video game playing
[11:20] Ian's career progression
[17:55] Lessons to apply in the Data space
[24:23] Challenges at Riot
[34:18] Real-time element
[39:09] Ian's day-to-day responsibilities
[43:13] Analysis vs. Production Code Quality
[48:11] Tools and techniques on the reality of writing production codes
[55:00] What would you change your career into?
[57:00] Ian's best practices advise
[58:28] Ian's favorite video game
[59:58] Wrap-up
MLOps Coffee Sessions #131 {Podcast BTS} with Ethan Rosenthal, Let's Continue Bundling into the Database co-hosted by Mike Del Balso.
// Abstract
The relationship between ML Engineers and Product Managers is something that we don't talk about enough. We've got to get this right. If we don't get this right, either you're not focusing on the business problems in the right way or the Product Managers are not going to understand the tech appropriately to help make the right decisions.
// Bio
Ethan works on the Conversations Team at Square leading a team of Artificial Intelligence Engineers. Ethan's team builds applied AI solutions for Square Messages, a messaging hub for Square merchants to communicate with their customers. Prior to Square, Ethan spent time as a freelance data science consultant building machine learning products for a range of companies, from pre-seed startups to Fortune 100 enterprises.
Ethan got his start in data science working at two different e-commerce startups, Birchbox and Dia&Co. Before data science, Ethan was an actual scientist and got his Ph.D. in experimental physics from Columbia University.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://www.ethanrosenthal.com/
Relevant blog posts:
https://www.ethanrosenthal.com/2022/05/10/database-bundling/
https://www.ethanrosenthal.com/2022/07/19/materialize-ml-monitoring/
https://www.ethanrosenthal.com/2022/01/18/autoretraining-is-easy/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mike on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/
Connect with Ethan on LinkedIn: https://www.linkedin.com/in/ethanrosenthal/
Timestamps:
[00:00] Ethan's preferred coffee
[00:10] Introduction to co-host Mike Del Balso
[00:43] Takeaways
[04:10] Sign up for our newsletter!
[04:47] Ethan's team
[06:49] Ethan's team improvement
[08:40] Product manager role at Square
[10:39] Large Language Models
[12:22] Big questions to figure out
[15:45] Cost of false-positive
[18:20] Company Vocabulary
[20:11] MLOps concerns and challenges around Large Language Models
[23:42] Data learning management
[27:36] Leveling trade-offs
[30:25] Ethan's Database Bundling blog
[34:32] Feature Stores
[38:24] Streaming databases
[41:57] Best of both worlds trade-off highlight
[43:51] Rosenthal data
[46:40] Ethan's freelancing
[47:46] Risk-reward trade-off
[49:17] Ethan as a professor
[51:14] Wrap up
MLOps Coffee Sessions #130 {Podcast BTS} with Andrew Yates, Adversarial MLOps on Other People's Money: Lessons Learned from Ad Tech co-hosted by Abi Aryan.
// Abstract
Design ML to be components in a larger system with stable interfaces is not tracible to monitor the entire stack as a black box. You need intermediate ground-truth signals. Ads are designed in this way.
You can go from profitable to non-profitable real quick with ads. This will determine whether your company is around a year or two. You play with money and sometimes you play a lot of it so make sure that it's correct.
// Bio
Andrew Yates formerly led ads ranking, auction, and marketplace engineering and research teams at Facebook and Pinterest. He specializes in designing billion-dollar content marketplaces that maximize long-term revenue while protecting both seller and user experiences. Andrew has published over a dozen patents in online advertising optimization.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Andrew on LinkedIn: https://www.linkedin.com/in/andrew-yates-0217a985/
MLOps Coffee Sessions #129 {Podcast BTS} with Catherin Breslin, Voice and Language Tech co-hosted by Adam Sroka.
// Abstract
Back in the day, Speech Recognition was its own thing. It's a very different flavor of Data Science. You could not use a lot of the tools. It wouldn't cross over to this type of machine learning.
Now, with the advancements, Speech Recognition and Machine learning are coming in together. It's interesting to hear right from someone with a Ph.D. level working with some of the biggest companies in the world doing it. The fact that something like Alexa is lots of models back to back and just fathom the complexity of that is quite cool!
// Bio
Catherine is a machine learning scientist and consultant based in Cambridge UK, and the founder of Kingfisher Labs consulting. Since completing her Ph.D. at the University of Cambridge in 2008, Catherine has commercial and academic experience in automatic speech recognition, natural language understanding, and human-computer dialogue systems, having previously worked at Cambridge University, Toshiba Research, Amazon Alexa, and Cobalt Speech. Catherine has been excited by the application of research to real-world problems involving speech and language at scale.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
www.catherinebreslin.co.uk
https://catherinebreslin.medium.com/
MLOps Community Newsletter: https://airtable.com/shrx9X19pGTWa7U3YTwitter: https://twitter.com/catherinebuk
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka
Connect with Catherine on LinkedIn: https://www.linkedin.com/in/catherine-breslin-0592423a/
Timestamps:
[00:00] Catherine's preferred coffee
[01:50] Takeaways
[03:59] Introduction to Catherine Breslin
[05:04] Subscribe to our newsletter!
[06:25] Catherine's background
[08:13] Speech Recognition trajectory
[09:36] Challenges around technologies and tools
[11:34] Reflective trend
[13:02] Developer experiences hiccups
[15:09] Speech Recognition use case backup
[16:56] Toshiba research
[17:48] Transition from a research lab to working in the industry
[20:01] Unit test of Speech Recognition
[20:56] Alexa
[22:33] Maturity process of Speech Recognition
[26:48] Speech Recognition unrecognizing challenges
[30:38] Mechanical Terk
[33:00] Social media listening
[34:05] Pipeline models and speed of Speech Recognition
[36:48] Development of Speech Recognition excited about
[37:23] Data from people for the Speech Recognition system vs Scowering news vs watching Youtube for a long time
[40:00] Disappearing Languages
[41:30] Future of an online practice partner
[43:17] Speech-to-speech translation
[44:04] Interesting ways to use unfamiliar models to achieve a result
[45:40] Meeting transcriptions
[48:37] First toy problem of a new Speech Recognition learner
[51:37] Kingfisher Labs' problems to tackle
[52:18] Off-the-shelf solution
[53:38] Translation layer
[54:15] Connect with Catherine on Twitter and LinkedIn for available jobs
[54:43] Wrap up
MLOps Coffee Sessions #128 with Simon Thompson, Managing Machine Learning Projects co-hosted by Abi Aryan.
// Abstract
It's a cliche to say that choosing and running the algorithms is only a small part of a typical ML project but despite that it's true! Setting up and organizing the project, dealing with the data asset, getting to the heart of the business problem, assessing and choosing the models, and integrating them with the business processes in production are all at least as time-consuming and important.
Simon has written a book that talks about how these different activities need to be orchestrated and executed and he hopes that it might be useful for people who are starting out managing ML projects and help them avoid some of the crunches and catches that seem to trip people up.
// Bio
Simon has been building and running ML projects since 1994 (when he started his Ph.D. in MachineLearning). His first commercial project was for the Royal Navy, and since then he has worked in Telecom, Defense, Consultancy, Manufacturing, and Finance. This means Simon has experienced a wide range of working environments and different types of projects. As well as working in a variety of commercial environments Simon collaborated on EU research projects, UK Government funded research projects and worked as an industrial rep on three MIT consortia (BigData@CSAIL, Systems That Learn, and the CISR Data Research Board).
Simon was also an industrial fellow at the Alan Turing Institute for a year. This means that he has also seen a lot of the communities' practices and concerns as they developed, and he had the chance to put them into use in a commercial environment.
Right now, Simon is working for a technology consultancy called GFT, and his job there is primarily to deliver ML projects for companies in the capital markets such as investment banks, although we also do work in retail banking, insurance, and manufacturing.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://medium.com/@sgt101
Managing Machine Learning Projects From design to deployment book by Simon Thompson:
https://www.manning.com/books/managing-machine-learning-projects
MLOps Community Newsletter: https://airtable.com/shrx9X19pGTWa7U3Y
Language processing. Simon Thompson CO545 Lecture 10: https://docplayer.net/211236676-Language-processing-simon-thompson-co545-lecture-10.html
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Abi on LinkedIn: https://www.linkedin.com/in/abiaryan/
Connect with Simon on LinkedIn: https://www.linkedin.com/in/simon-thompson-025a7/
MLOps Coffee Sessions #127 with Niall Murphy & Todd Underwood, Reliable ML co-hosted by David Aponte.
// Abstract
By applying an SRE mindset to machine learning, authors and engineering professionals Cathy Chen, Kranti Parisa, Niall Richard Murphy, D. Sculley, Todd Underwood, and featured guest authors show you how to run an efficient and reliable ML system. Whether you want to increase revenue, optimize decision-making, solve problems, or understand and influence customer behavior, you'll learn how to perform day-to-day ML tasks while keeping the bigger picture in mind. (Book description from O'Reilly)
MLOps Coffee Sessions #127 with Niall Murphy & Todd Underwood, Reliable ML co-hosted by David Aponte. // Abstract By applying an SRE mindset to machine learning, authors and engineering professionals Cathy Chen, Kranti Parisa, Niall Richard Murphy, D. Sculley, Todd Underwood, and featured guest authors show you how to run an efficient and reliable ML system. Whether you want to increase revenue, optimize decision-making, solve problems, or understand and influence customer behavior, you'll learn how to perform day-to-day ML tasks while keeping the bigger picture in mind. (Book description from O'Reilly) It was great that they wrote this book in the first place in a space that's new and lots of people are entering it with a lot of questions and this book clarifies those questions. It was also great to have all of their experiences documented in this one book and there's a lot of value in putting them all in one place so that people can benefit from it.
// Bio
Niall Murphy
Niall has been interested in Internet infrastructure since the mid-1990s. He has worked with all of the major cloud providers from their Dublin, Ireland offices - most recently at Microsoft, where he was the global head of Azure Site Reliability Engineering (SRE). His books have sold approximately a quarter of a million copies worldwide, most notably the award-winning Site Reliability Engineering, and he is probably one of the few people in the world to hold degrees in Computer Science, Mathematics, and Poetry Studies. He lives in Dublin, Ireland, with his wife and two children.
Todd Underwood
Todd is a Director at Google and leads Machine Learning for Site Reliability Engineering Director. He is also the Site Lead for Google’s Pittsburgh office. ML SRE teams build and scale internal and external ML services and are critical to almost every Product Area at Google.
Before working at Google, Todd held a variety of roles at Renesys. He was in charge of operations, security, and peering for Renesys’s Internet intelligence services which are now part of Oracle's Cloud service. He also did product work for some early social products that Renesys worked on. Before that Todd was the Chief Technology Officer of Oso Grande, an independent Internet service provider (AS2901) in New Mexico.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Reliable Machine Learning book by Cathy Chen, Niall Richard Murphy, Kranti Parisa, D. Sculley, Todd Underwood: https://www.oreilly.com/library/view/reliable-machine-learning/9781098106218/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Niall on LinkedIn: https://www.linkedin.com/in/niallm/
Connect with Todd on LinkedIn: https://www.linkedin.com/in/toddunder/
MLOps Coffee Sessions #126 with George Mathew, ML Unicorn Start-up Investor Tells-IT-All.
// Abstract
What's so enticing about enterprise software? It's incredible to see George's idea and vision to invest in generationally enduring companies.
Let's look at the way how George likes to structure deals with companies while he's reviewing them and let's look at the MLOps ecosystem through the eyes of the investors.
// Bio
George Mathew joins Insight Partners as a Managing Director focused on venture stage investments in AI, ML, Analytics, and Data companies as they are establishing product/market Fit.
He brings 20+ years of experience developing high-growth technology startups including most recently being CEO of Kespry. Prior to Kespry, George was President & COO of Alteryx where he scaled the company through it’s IPO (AYX). Previously he held senior leadership positions at SAP and salesforce.com. He has driven company strategy, led product management and development, and built sales and marketing teams.
George holds a Bachelor of Science in Neurobiology from Cornell University and a Masters in Business Administration from Duke University, where he was a Fuqua Scholar.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://www.insightpartners.com/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with George on LinkedIn: https://www.linkedin.com/in/gmathew/
MLOps Coffee Sessions #125 with Rafael Pierre, Deploying Real-time ML Models in Minutes with Databricks Model Serving V2 co-hosted by Ryan Russon.
// Abstract
From our experience helping customers in the Data and AI field, we learned that the most challenging part of Machine Learning is deploying it. Putting models into production is complex and requires additional pieces of infrastructure as well as specialized people to take care of it - this is especially true if we are talking about real-time REST APIs for serving ML models.
With Databricks Model Serving V2, we introduce the idea of Serverless REST endpoints to the platform. This allows teams to easily deploy their ML models in a production-grade platform with a few mouse clicks (or lines of code 😀).
// Bio
Rafael has worked for 15 years in data-intensive fields within finance in multiple roles: software engineering, product management, data engineering, data science, and machine learning engineering.
At Databricks, Rafael has fun bringing all these topics together as a Solutions Architect to help our customers become more and more data-driven.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://mlopshowto.com
Airflow Summit 2022:
https://youtu.be/JsYEOdRBgREING
Data Engineering Meetup:
https://www.youtube.com/watch?v=gJoxX1rRZJI
MLOps World Virtual Summit NYC 2022:
https://drive.google.com/file/d/1EXsqmLfrPAsV9i6h6pGfJxVjMO9y6u9a/view?usp=sharing
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Ryan on LinkedIn: https://www.linkedin.com/in/ryanrusson/
Connect with Rafael on LinkedIn: https://www.linkedin.com/in/rafaelpierre
Lightning Sessions #2 with Aparna Dhinakaran, Co-Founder and Chief Product Officer, and Jason Lopatecki, CEO and Co-Founder of Arize. Lightning Sessions is sponsored by Arize
// Abstract
Monitoring embeddings on unstructured data is not an easy feat let's be honest. Most of us know what it is but don't understand it one hundred percent.
Thanks to Aparna and Jason of Arize for breaking down embedding so clearly. At the end of this Lightning talk, we get to see a demo of how Arize deals with unstructured data and how you can use Arize to combat that.
// Bio
Aparna Dhinakaran
Aparna is the Co-Founder and Chief Product Officer at Arize AI, a pioneer, and early leader in machine learning (ML) observability. A frequent speaker at top conferences and thought leader in the space, Dhinakaran was recently named to the Forbes 30 Under 30. Before Arize, Dhinakaran was an ML engineer and leader at Uber, Apple, and TubeMogul (acquired by Adobe). During her time at Uber, she built several core ML Infrastructure platforms, including Michaelangelo.
Aparna has a BA from Berkeley's Electrical Engineering and Computer Science program, where she published research with Berkeley's AI Research group. She is on a leave of absence from the Computer Vision Ph.D. program at Cornell University.
Jason Lopatecki
Jason is the Co-founder and CEO of Arize AI, a machine learning observability company. He is a garage-to-IPO executive with an extensive background in building marketing-leading products and businesses that heavily leverage analytics. Prior to Arize, Jason was co-founder and chief innovation officer at TubeMogul where he scaled the business into a public company and eventual acquisition by Adobe.
Jason has hands-on knowledge of big data architectures, programmatic advertising systems, distributed systems, and machine learning and data processing architectures. In his free time, Jason tinkers with personal machine learning projects as a hobby, with a special interest in unsupervised learning and deep neural networks. He holds an electrical engineering and computer science degree from UC Berkeley.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
https://arize.com/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Aparna on LinkedIn: https://www.linkedin.com/in/aparnadhinakaran/
Connect with Jason on LinkedIn: https://www.linkedin.com/in/jason-lopatecki-9509941/
Timestamps:
[00:00] Introduction to the topic
[01:13] Troubleshooting unstructured ML models is difficult
[01:40] Challenges with monitoring unstructured data
[02:10] How data looks like
[03:02] Embeddings are the backbone of unstructured models
[03:28] ML teams need a common tool
[04:06] What are embeddings?
[05:08] The real WHY behind AI
[06:41] ML observability for unstructured data
[07:08] Index and Monitor every Embedding
[08:05] Measuring drift of unstructured data
[08:54] Interactive visualizations
[09:34] Fix underlying data issue
[09:44] Data-centric AI workflow
[10:08] Demo of the product
[12:48] Wrap up
MLOps Coffee Sessions #124 with Kush Varshney, Distinguished Research Staff Member and Manager IBM Research, Trustworthy Machine Learning co-hosted by Krishnaram Kenthapadi.
// Abstract
Trustworthy ML is a way of thinking and something to be worked on and operationalized throughout the entire machine learning development lifecycle, starting from the problem specification phase that should include diverse stakeholders.
// Bio
Kush R. Varshney was born in Syracuse, New York in 1982. He received a B.S. degree (magna cum laude) in electrical and computer engineering with honors from Cornell University, Ithaca, New York, in 2004. He received the S.M. degree in 2006 and the Ph.D. degree in 2010, both in electrical engineering and computer science at the Massachusetts Institute of Technology (MIT), Cambridge. While at MIT, he was a National Science Foundation Graduate Research Fellow.
Dr. Varshney is a distinguished research staff member and manager with IBM Research at the Thomas J. Watson Research Center, Yorktown Heights, NY, where he leads the machine learning group in the Foundations of Trustworthy AI department. He was a visiting scientist at IBM Research - Africa, Nairobi, Kenya in 2019. He is the founding co-director of the IBM Science for Social Good initiative. He applies data science and predictive analytics to human capital management, healthcare, olfaction, computational creativity, public affairs, international development, and algorithmic fairness, which has led to recognitions such as the 2013 Gerstner Award for Client Excellence for contributions to the WellPoint team and the Extraordinary IBM Research Technical Accomplishment for contributions to workforce innovation and enterprise transformation. He conducts academic research on the theory and methods of trustworthy machine learning. His work has been recognized through best paper awards at the Fusion 2009, SOLI 2013, KDD 2014, and SDM 2015 conferences and the 2019 Computing Community Consortium / Schmidt Futures Computer Science for Social Good White Paper Competition. He self-published a book entitled 'Trustworthy Machine Learning in 2022, available at http://www.trustworthymachinelearning.com. He is a senior member of the IEEE.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Krishnaram on LinkedIn: https://www.linkedin.com/in/krishnaramkenthapadi
Connect with Kush on LinkedIn: https://www.linkedin.com/in/kushvarshney/
MLOps Coffee Sessions #123 with Gleb Abroskin, Machine Learning Engineer at Funcorp, RECOMMENDER SYSTEM: Why They Update Models 100 Times a Day co-hosted by Jake Noble.
// Abstract
FunCorp was a top 10 app store. It was a very popular app that has a ton of downloads and just memes. They need a recommendation system on top of that. Memes are super tricky because they're user-generated and they evolve very quickly. They're going to live and die by the Recommender System in that product.
It's incredible to see FunCorp's maturity. Gleb breaks down the feature store they created and the velocity they have to be able to create a whole new pipeline in a new model and put it into production after only a month!
// Bio
Gleb make models go brrrrr. He doesn't know what is expected in this field, to be honest, but Gleb has experience in deploying a lot of different ML models for CV, speech recognition, and RecSys in a variety of languages (C++, Python, Kotlin) serving millions of users worldwide.
/ MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Putting a two-layered recommendation system into production -
https://medium.com/@FunCorp/putting-a-two-layered-recommendation-system-into-production-b8caaf61393d
Practical Guide to Create a Two-Layered Recommendation System -
https://medium.com/@FunCorp/practical-guide-to-create-a-two-layered-recommendation-system-5486b42f9f63
Ten Mistakes to Avoid When Creating a Recommendation System -
https://medium.com/@FunCorp/ten-mistakes-to-avoid-when-creating-a-recommendation-system-8268ed60aeba
Applying Domain-Driven Design And Patterns: With Examples in C# and .net 1st Edition by Jimmy Nilsson:
https://www.amazon.com/Applying-Domain-Driven-Design-Patterns-Examples/dp/0321268202
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Jake on LinkedIn: https://www.linkedin.com/in/jakednoble/
Connect with Gleb on LinkedIn: https://www.linkedin.com/in/gasabr/
Timestamps:
[00:00] Introduction to Gleb Abroskin
[00:50] Takeaways
[05:39] Breakdown of FunCorp teams
[06:47] FunCorp's team ratio
[07:41] FunCorp team provisions
[08:48] Feature Store vision
[10:16] Matrix factorization
[11:51] Fairly modular fairly thin infrastructure
[12:26] Distinct models with the same feature
[13:08] FunCorp's definition of Feature Store
[15:10] Unified API
[15:55] FunCorp's scaling direction
[17:01] Level up as needed
[17:38] Future of FunCorp's Feature Store
[18:37] Monitoring investment in the space
[19:43] Latency for business metrics
[21:04] Velocity to production
[23:10] 30-day retention struggle
[24:45] Back-end business stability
[27:49] Recommender systems
[30:34] Back-end layer headaches
[32:04] Missing piece of the whole Feature Store picture
[33:54] Throwing ideas turn around time
[36:37] Decrease time to market
[37:41] Continuous training pipelines or produce an artifact
[39:33] Worst-case scenario
[40:38] Realistic estimation of a new model deployment
[41:42] Recommender Systems' future velocity
[43:07] A/B Testing launch - no launch decision
[46:32] Lightning question
[47:08] Wrap up
MLOps Coffee Sessions #122 with Hannes Hapke, Machine Learning Engineer at Digits Financial, Inc., Scaling Similarity Learning at Digits co-hosted by Vishnu Rachakonda.
// Abstract
Machine Learning in a product is a double-edged sword. It can make a product more useful but it depends on assumed and strictly defined behavior from users.
Hannes walks through the entirety of their machine learning pipeline, how they implemented it, what the elements are, what the learning looks like, and what tooling looks like.
Hannes maps out what good data hygiene looks like not only from the machine learning perspective down to the software engineering, design, and backend engineering, all the way to the data engineering perspectives.
// Bio
Hannes was the first ML engineer at Digits, where he built the MLOPs foundation for their ML team. His interest in production machine learning ranges from building ML pipelines to scaling similarity-based ML to process millions of banking transactions daily.
Prior to Digits, Hannes implemented ML solutions for a number of applications, incl. retail, health care, or ERP companies.
He co-author two machine learning books:
* Building Machine Learning Pipeline (O'Reilly)
* NLP in Action (Manning)
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Hannes on LinkedIn: https://www.linkedin.com/in/hanneshapke/
Timestamps:
[00:00] Introduction to Hannes Hapke
[01:37] Takeaways
[02:40] Design supercharges machine learning
[05:48] Building Machine Learning Pipeline book
[08:09] Updating the edition
[09:37] Abstract away
[11:52] Approach of crossover
[16:04] Training serving skew
[20:42] Tools using continuous integration and deployment
[25:25] Human in the loop touch point
[27:44] Data backfilling update
[30:06] Work and Products of Digits
[32:26] Digit Boost
[35:30] The first machine learning engineer
[39:55] Structured data in good shape, good data processing perspective, concept-educated teams
[43:33] Digits is hiring!
[43:55] Machine Learning struggles
[47:10] Design decision
[49:49] Data or machine learning literacy
[51:30] Data Hygiene
[52:49] Rapid fire questions
[54:47] Wrap up
MLOps Coffee Sessions #121 with Luis Ceze, CEO and Co-founder of OctoML, Bringing DevOps Agility to ML co-hosted by Mihail Eric.
// Abstract
There's something about this idea where people see a future where you don't need to think about infrastructure. You should just be able to do what you do and infrastructure happens.
People understand that there is a lot of complexity underneath the hood and most data scientists or machine learning engineers start deploying things and shouldn't have to worry about the most efficient way of doing this.
// Bio
Luis Ceze is Co-Founder and CEO of OctoML, which enables businesses to seamlessly deploy ML models to production making the most out of the hardware. OctoML is backed by Tiger Global, Addition, Amplify Partners, and Madrona Venture Group. Ceze is the Lazowska Professor in the Paul G. Allen School of Computer Science and Engineering at the University of Washington, where he has taught for 15 years.
Luis co-directs the Systems and Architectures for Machine Learning lab (sampl.ai), which co-authored Apache TVM, a leading open-source ML stack for performance and portability that is used in widely deployed AI applications.
Luis is also co-director of the Molecular Information Systems Lab (misl.bio), which led pioneering research in the intersection of computing and biology for IT applications such as DNA data storage. His research has been featured prominently in the media including New York Times, Popular Science, MIT Technology Review, and the Wall Street Journal. Ceze is a Venture Partner at Madrona Venture Group and leads their technical advisory board.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Landing page: https://octoml.ai/
The Boys in the Boat: Nine Americans and Their Epic Quest for Gold at the 1936 Berlin Olympics by Daniel James Brown:
https://www.amazon.com/Boys-Boat-Americans-Berlin-Olympics/dp/0143125478
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Luis on LinkedIn: https://www.linkedin.com/in/luis-ceze-50b2314/
Timestamps:
[00:00] Introduction to Luis Ceze
[06:28] MLOps does not exist
[10:41] Semantics argument
[16:25] Parallel programming standpoint
[18:09] TVM
[22:51] Optimizations
[24:18] TVM in the ecosystem
[27:10] OctoML's further step
[30:42] Value chain
[33:58] Mature players
[35:48] Talking to SRE's and Machine Learning Engineers
[36:32] Building OctoML
[40:20] My Octopus Teacher
[42:15] Environmental effects of Sustainable Machine Learning
[44:50] Bridging the gap from OctoML to biological mechanisms
[50:02] Programmability
[57:13] Academia making the impact
[59:40] Rapid fire questions
[1:03:39] Wrap up
MLOps Coffee Sessions #120 with David Stein, Senior Staff Software Engineer at LinkedIn, Feathr: LinkedIn's Enterprise-Grade, High-Performance Feature Store co-hosted by Skylar Payne.
// Abstract
When David started building Feathr, Feature Stores did not exist. That was not a term floating around at all. This was definitely one of the OG Feature Stores for sure!
We hear how the LinkedIn team got to this point, having an open source release, and how they used LinkedIn as an incubator to build a great product.
// Bio
David Stein is a tech lead at LinkedIn working on machine learning feature infrastructure. He is the original architect of Feathr and continues to contribute to its development as well as to other parts of LinkedIn's machine learning platform.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://engineering.linkedin.com/blog/2022/open-sourcing-feathr---linkedin-s-feature-store-for-productive-m
* https://github.com/linkedin/feathr
* https://engineering.linkedin.com/blog/2022/open-sourcing-feathr---linkedin-s-feature-store-for-productive-m
* https://azure.microsoft.com/en-us/blog/feathr-linkedin-s-feature-store-is-now-available-on-azure/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with David on LinkedIn: https://www.linkedin.com/in/steindavidj/
Timestamps:
[00:00] Introduction to David Stein
[00:30] Takeaways
[06:58] David and Skylar's reunion
[08:34] Feathr's background
[12:04] Lessons learned in building Feathr
[16:35] Scale of Feathr
[20:25] Systems interaction
[24:15] Standardization
[29:20] Importance and better difference of Feathr
[34:30] Feature Stores' evolution and more generally MLOps
[37:57] Challenges in real-time
[40:09] Going real-time, you're not ready!
[42:24] Use cases strong for real-time unleveraged now
[45:50] Proud about
[47:58] Going back in time
[50:47] Feathr as a separate company?
[52:53] LinkedIn is hiring!
MLOps Coffee Sessions #119 with Hien Luu, Sr. Engineering Manager of DoorDash, MLOps at DoorDash: 3 Principles for Building an ML Platform That Will Sustain Hypergrowth co-hosted by Skylar Payne.
// Abstract
Machine Learning plays a big part at DoorDash in terms of what they do on a daily basis. It powers many of their core infrastructures.
When it comes to DoorDash's business, they have to be leveraging machine learning and it is such a huge piece of the business that it is critical.
// Bio
Hien Luu is an Engineering Manager at DoorDash, leading the Machine Learning platform team at DoorDash. He is particularly passionate about the intersection between Artificial Intelligence and Big Data. He is the author of the Beginning Apache Spark 3 book. He has given presentations at various conferences like Data+AI Summit, MLOps World, Deep Learning Summit, and apply() conference.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
engineering.linkedin.com/hadoop/user-engagement-powered-apache-pig-and-hadoop
* https://doordash.engineering/2020/07/20/enabling-efficient-machine-learning-model-serving/
* https://doordash.engineering/2020/11/19/building-a-gigascale-ml-feature-store-with-redis/
* https://doordash.engineering/2021/03/04/building-a-declarative-real-time-feature-engineering-framework/
* https://doordash.engineering/2021/05/20/monitor-machine-learning-model-drift/
* https://doordash.engineering/2021/01/26/computational-graph-machine-learning-ensemble-model-support/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Hien on LinkedIn: https://www.linkedin.com/in/hienluu/
Timestamps:
[00:00] Introduction of DoorDash team
[01:58] Overview of DoorDash
[03:32] DoorDash's platform
[13:23] Experimenting and testing new models
[15:15] Experience transferring
[17:16] Effective engagement with customers
[24:15] Team sizes
[25:37] Metrics
[33:25] App for users
[34:04] Using Databricks and Snowflake together
[37:49] Supporting power users
[40:17] Advice and experiences
[43:53] Wrap up
MLOps Coffee Sessions #118 with Olalekan Elesin, Director of Data Platform & Data Architect at HRS Product Solutions GmbH, co-hosted by Vishnu Rachkonda.
// Abstract
You don't have infinite resources? Call out your main metrics! Focus on the most impactful things that you could do for your data scientists. Olalekan joined us to talk about his experience previously building a machine learning platform at Scaleout24.
From our standpoint, this is the best demonstration and explanation of the role of technical product management in ML that we have on the podcast so far!
// Bio
Olalekan Elesin is a technologist with a successful track record of delivering data-driven technology solutions that leverages analytics, machine learning, and artificial intelligence. He combines experience working across 2 continents and 5 different market segments ranging from telecommunications, e-commerce, online marketplaces, and current business travel.
Olalekan built the AI Platform 1.0 at Scout24 and currently leads multiple data teams at HRS Group. He is an AWS Machine Learning Community Hero in his spare time.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
What Customers Want: Using Outcome-Driven Innovation to Create Breakthrough Products and Services book by Anthony Ulwick: https://www.amazon.com/What-Customers-Want-Outcome-Driven-Breakthrough/dp/0071408673
Empowered: Ordinary People, Extraordinary Products by Marty Cagan:
https://www.amazon.com/EMPOWERED-Ordinary-Extraordinary-Products-Silicon/dp/111969129X
How to Avoid a Climate Disaster: The Solutions We Have and the Breakthroughs We Need by Bill Gates:
https://www.amazon.com/How-Avoid-Climate-Disaster-Breakthroughs/dp/059321577X
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Olalekan on LinkedIn: https://www.linkedin.com/in/elesinolalekan/
Timestamps:
[00:00] Introduction to Olalekan Elesin
[00:42] Takeaways
[02:52] Situation at Scaleout24
[07:53] Data landscape engineer and architect
[11:27] Depiction of events
[13:53] Platform approach investment
[15:59] Exceptional need or opportunity to the most intense need
[17:41] Long-tail pieces
[22:01] Metrics
[24:15] Nitty-gritty product works
[26:00] Educating people metrics
[30:02] Upskilling fundamentals of the product discipline
[34:05] Investing in AWS
[37:53] Best-of-breed tools
[44:34] Continuous development for AutoML
[47:26] Rapid fire questions
[52:19] Wrap up
MLOps Coffee Sessions #117 with Chad Sanderson, Head of Product, Data Platform at Convoy, Data Engineering for ML co-hosted by Josh Wills.
// Abstract
Data modeling is building relationships between core concepts within your data. The physical data model shows how the relationships manifest in your data environment but then there's the semantic data model, the way that entity relationship design is extracted away from any data-centric implementation.
Let's do the good old fun of talking about why data modeling is so important!
// Bio
Chad Sanderson is the Product Lead for Convoy's Data Platform team, which includes the data warehouse, streaming, BI & visualization, experimentation, machine learning, and data discovery.
Chad has built everything from feature stores, experimentation platforms, metrics layers, streaming platforms, analytics tools, data discovery systems, and workflow development platforms. He’s implemented open source, SaaS products (early and late-stage) and has built cutting-edge technology from the ground up. Chad loves the data space, and if you're interested in chatting about it with him, don't hesitate to reach out.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://odsc.com/speakers/scaling-machine-learning-with-data-mesh/ https://docs.google.com/presentation/d/1rVtltHkRkP_JaGZdkAS3U_SXfr5Gg-RP980FKXh0YNU/edit?usp=sharing
Josh Wills will be teaching a course on Data Engineering for Machine Learning in September here:
https://www.getsphere.com/ml-engineering/data-engineering-for-machine-learning
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Josh on LinkedIn: https://www.linkedin.com/in/josh-wills-13882b/
Connect with Chad on LinkedIn: https://www.linkedin.com/in/chad-sanderson/
Timestamps:
[00:00] Introduction of the new co-host Josh Wills
[00:54] Introduction to Chad Sanderson
[01:46] Josh will lead a course for Machine Learning in mid-September
[02:16] Data modeling blog post of Chad
[06:10] Idea of Strategy
[09:40] Modern cloud data warehouses
[17:01] Layering on contracts
[20:38] Scaling at larger companies
[25:30] Carrot-stick strategy
[34:27] Second and third-order effects
[39:53] Stockholm Syndrome
[41:22] Quality checks at Slack
[45:28] Success in two main ways according to Chad
[47:35] Completely and utterly different universes
[53:42] Product use case to push semantic events
[56:00] Pattern of analysis of the sequence of events
[57:23] Wrap up
MLOps Coffee Sessions #116 with Shawn Kyzer, Principal Data Engineer at Thoughtworks (Spain), Scaling Machine Learning with Data Mesh co-hosted by Adam Sroka.
// Abstract
You can't just get something done by using tools. You need to go much deeper than that and it is very clear that Data Mesh is the same thing. You have to educate the organization about the movement.
In this session, Shawn broke down the cultural piece of data mesh and how many parallels there are with the MLOps Movement when it comes to the cultural side of MLOps.
// Bio
Shawn is passionate about harnessing the power of data strategy, engineering, and analytics in order to help businesses uncover new opportunities. As an innovative technologist with over 13 years of experience, Shawn removes technology as a barrier and broadens the art of the possible for business and product leaders. His holistic view of technology and emphasis on developing and motivating strong engineering talent, with a focus on delivering outcomes whilst minimising outputs, is one of the characteristics which sets him apart from the crowd.
Shawn’s deep technical knowledge includes distributed computing, cloud architecture, data science, machine learning, and engineering analytics platforms. He has years of experience working as a consultant practitioner for a variety of prestigious clients ranging from secret clearance level government organizations to Fortune 500 companies.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://odsc.com/speakers/scaling-machine-learning-with-data-mesh/ https://docs.google.com/presentation/d/1rVtltHkRkP_JaGZdkAS3U_SXfr5Gg-RP980FKXh0YNU/edit?usp=sharing
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Shawn on LinkedIn: https://www.linkedin.com/in/shawn-kyzer-msit-mba-b5b8a4b/
Timestamps:
[00:00] Introduction to Shawn Kyzer
[00:43] Takeaways
[04:00] Data Mesh for ML projects
[11:22] The signal for the exploratory part of a new modeling project
[14:13] Ownership and centralization
[16:20] Lack of technology and some implementations literature
[17:10] Python stronghold from Microsoft blogs
[23:09] Integration with self-serve data platform
[25:31] Starting a platform team
[30:04] Quick wins
[32:09] Metrics monitoring
[34:18] Metrics break up
[38:32] Limit to capabilities and not worth doing
[41:39] Culture and technology holds
[44:03] Setting the foundation
[46:53] Unforeseen benefits
[52:19] Lightning question
MLOps Coffee Sessions #115 with Flaviu Vadan, Senior Software Engineer at Dyno Therapeutics, How Hera is an Enabler of MLOps Integrations co-hosted by Vishnu Rachakonda.
// Abstract
Flaviu talks about the internal ML platform at Dyno Therapeutics called Hera. His team uses Hera as an internal innovation engine to help discover new breakthroughs with machine learning in the biotech healthcare industry.
/ Bio
Flaviu is a Senior Software Engineer at Dyno Therapeutics, the leading organization in the design of novel gene therapy vectors with transformative delivery properties for a vast landscape of human diseases. Flaviu comes from a background focused on Bioinformatics, which is a field that combines Computer Science, Mathematics, and Biology. He took stints in academia by working as a research assistant in Computer Science and Bioinformatics labs before joining Dyno Therapeutics to work on machine-guided design of adeno-associated viruses (AAVs).
At Dyno, Flaviu works on compute and core infrastructure, DevOps, MLOps, and approaches that combine AI/ML to design AAVs in silico. He is also the author and maintainer of Hera, a Python SDK that facilitates access to Argo Workflows by making workflow construction and submission easy and accessible.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Flaviu on LinkedIn: https://www.linkedin.com/in/flaviuvadan/
Timestamps:
[00:00] Introduction to Flaviu Vadan
[00:50] Takeaways
[02:06] Share this episode with a friend!
[03:20] What Dyno does
[05:44] CRISPR and Gene Editing
[06:21] Kidney transplants and using pig organs
[07:31] Deciding what genes to put in the body
[07:48] Role of ML at Dyno
[10:07] Higher dose
[13:41] Process of Machine Learning Deployment and Productionizing at Dyno
[16:22] Proliferation of models
[17:31] Building the internal platform
[19:37] Interaction with data, translation to compute layer, evaluation
[24:21] Venn diagram for MLOps
[27:06] Leveraging Argo Workflows
[30:34] Hera
[35:28] Open sourcing
[38:44] Human power at Dyno
[41:17] Wrap up
MLOps Coffee Sessions #114 with Marc Lindner, Co-Founder COO and Amr Mashlah, Head of Data Science of eezylife Inc., Product Enrichment and Recommender Systems co-hosted by Skylar Payne.
// Abstract
The difficulties of making multi-modal recommender systems. How it can be easy to know something about a user but very hard to know the same thing about a product and vice versa? For example, you can clearly know that a user wants an intellectual movie, but it is hard to accurately classify a movie as intellectual and fully automated.
// Bio
Marc Lindner Marc has a background in Knowledge Engineering. He's Always extremely product-focused with anything to do with Machine Learning.
Marc built several products working together with companies such as Lithium Technologies etc. and then co-Founded eezy.
Amr Mashlah
Amr is the head of data science at eezy, where he leads the development of their recommender engine. Amr has a master's degree in AI and has been working with startups for 6 years now.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Children of Time book by Adrian Tchaikovsky:
https://www.amazon.com/Children-Time-Adrian-Tchaikovsky/dp/0316452505
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Marc on LinkedIn: https://www.linkedin.com/in/marc-lindner-883a0883/
Connect with Amr on LinkedIn: https://www.linkedin.com/in/mashlah/
MLOps Coffee Sessions #113 with Leanne Fitzpatrick, Director of Data Science of Financial Times, Building Better Data Teams co-hosted by Mihail Eric.
// Abstract
We spent a lot of time talking about data tooling but we maybe spent not as much time talking about data organizations and efficiently running and organizing data teams.
What about starting with limitations instead of aspirations? Right constraints instead of the north star? In this session, let's learn more about a realistic take on the state of data organizations of today.
// Bio
Leanne is Director of Data Science at the Financial Times and is a passionate data leader with experience building and developing empowered data science and analytics teams in a variety of businesses. Leanne is in her element when developing and implementing strategic, technical, and cultural solutions to getting machine learning and data science into the operational ecosystem.
Leanne is an active part of the data and technology community, sharing innovation and insights to encourage best practices, from Manchester, UK to Austin, TX, and is an Advisory Panel Board Member. Outside of all things data you can ask Leanne about her golf swing (it’s not good - yet), her passion for American Football (specifically the Cincinnati Bengals), her latest sewing project, and her love for good music, food, and whisky.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related
Links Children of Time book by Adrian Tchaikovsky:
https://www.amazon.com/Children-Time-Adrian-Tchaikovsky/dp/0316452505
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Leanne on LinkedIn: https://www.linkedin.com/in/leanne-kim-fitzpatrick-29204341/
Timestamps:
[00:00] Introduction to Leanne Fitzpatrick
[04:23] Write us your suggestions!
[05:43] Tri-pawed dog called Seaweed!
[08:43] How to architect data teams
[14:44] Organizational deficiencies
[19:19] Tensions and conflicts for starters
[24:07] Misunderstandings from marketing
[25:59] The Middle Layer
[28:48] Data science work at publications
[31:11] Mystique of going to real-time
[35:29] Third parties with fraud
[37:40] Augmenting data practitioners with third-party tools
[41:00] Principle of reinventing the wheel and avoiding undifferentiated heavy lifting
[46:29] Different Abstraction Layer recommendations
[48:42] RN Production
[51:56] Will Python eats RN Production away?
[56:05] Julia as a dark horse
[56:39] Future of RN Production
[58:00] Rapid fire questions
MLOps Coffee Sessions #112 with Xiangrui Meng, Principal Software Engineer of Databricks, MLX: Opinionated ML Pipelines in MLflow co-hosted by Vishnu Rachakonda.
// Abstract
MLX is to enable data scientists to stay mostly within their comfort zone utilizing their expert knowledge while following the best practices in ML development and delivering production-ready ML projects, with little help from production engineers and DevOps.
// Bio
Xiangrui Meng is a Principal Software Engineer at Databricks and an Apache Spark PMC member. His main interests center around simplifying the end-to-end user experience of building machine learning applications, from algorithms to platforms and to operations.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Good Strategy Bad Strategy: The Difference and Why It Matters book by Richard Rumelt:
https://www.amazon.com/Good-Strategy-Bad-Difference-Matters/dp/0307886239
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Xiangrui on LinkedIn: https://www.linkedin.com/in/mengxr/
Timestamps:
[00:00] Introduction to Xiangrui Meng
[00:39] Takeaways
[02:09] Xiangrui's background
[03:38] What kept Xiangrui in Databricks
[07:33] What needs to be done to get there
[09:20] Machine Learning passion of Xiangrui
[11:52] Changes in building that keep you fresh for the future
[14:35] Evolution core challenges to real-time and use cases in real-time
[17:33] DevOps + DataOps + ModelOps = MLOps
[19:21] MLFlow Support
[21:37] Notebooks to production debates
[25:42] Companies tackling Notebooks to production
[27:40] MLOoops stories
[31:03] Opinionated MLOps productionizing in a good way
[40:23] Xiangrui's MLOps Vision
[44:47] Lightning round
[48:45] Wrap up
MLOps Coffee Sessions #111 with Samuel Partee, Principal Applied AI Engineer of Redis, More than a Cache: Turning Redis into a Composable, ML Data Platform co-hosted by Mihail Eric. This episode is sponsored by Redis.
// Abstract
Pushing forward the Redis platform to be more than just the web-serving cache that we've known it up to now. It seems like a natural progression for the platform, we see how they're evolving to be this AI-focused, AI native serving platform that does vector similarity, feature stored provides those kinds of functionalities.
// Bio
A Principal Applied AI Engineer at Redis, Sam helps guide the development and direction of Redis as an online feature store and vector database.
Sam's background is in high-performance computing including ML-related topics such as distributed training, hyperparameter optimization, and scalable inference.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://partee.io
Redis VSS demo: https://github.com/Spartee/redis-vector-search
Redis Stack: https://redis.io/docs/stack/
Github - https://github.com/Spartee
OSS org Sam co-founded at HPE/Cray - https://github.com/CrayLabs
This paper last year was some of the best research and collaborations Sam has been a part of. The Paper is published here: https://www.sciencedirect.com/science/article/pii/S1877750322001065?via%3Dihub
Do you really need an extra database for vectors? https://databricks.com/dataaisummit/session/emerging-data-architectures-approaches-real-time-ai-using-redis
Blink: The Power of Thinking Without Thinking by Malcolm Gladwell, Barry Fox, Irina Henegar (Translator): https://www.goodreads.com/book/show/40102.Blink
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Sam on LinkedIn: www.linkedin.com/in/sam-partee-b04a1710a
Timestamps:
[00:00] Introduction to Samuel Partee
[00:24] Takeaways
[02:46] Updates on the Community
[05:17] Start of Redis
[08:10] Vision for Vector Search
[11:05] Changing the narrative going from the "Cache" for all servers and web endpoints
[14:35] Clear value prop on demos
[20:17] Vector Database
[26:26] Features with benefits
[28:41] AWS Spend
[30:39] Vector Database upsell model and bureaucratic convenience
[32:08] Distributed training hyperparameter optimization and scalable inference
[35:03] Core infrastructural advancement
[36:55] Tools movement to help
[39:00] Using Machine Learning at scale in numerical simulations with SmartSim: An application to ocean climate modeling (published paper) [42:52] Future applications of tech to get excited with
[44:20] Lightning round
[47:48] Wrap up
MLOps Coffee Sessions #110 with David Bayliss, Chief Data Scientist of LexisNexis Risk Solutions, Just Fetch the Data and then... co-hosted by Vishnu Rachakonda.
// Abstract
Composing data to extract features can be a significant problem. Key factors are the data size, compliance restrictions, and real-time data. Ethics (and law) can drive extremely complex audit requirements. In the cloud, you can do anything - at a price.
// Bio
One of the creators of the world's first big data platform (HPCC); David has been tackling big data problems for two decades. A mathematician, compiler writer, and data sponge with more than five dozen patents spanning platforms linking, and search.
Most inventors think outside the box; David can't even remember where the box is. He leads the team that creates their core Data Science methods used by hundreds of data scientists.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Interesting insight in this post. Would be cool to learn from David about his view on things
https://www.google.com/url?q=https://www.linkedin.com/posts/david-bayliss-426556a_datascience-platform-portability-activity-6913448643303759872-2dqq?utm_source%3Dlinkedin_share%26utm_medium%3Dmember_desktop_web&sa=D&source=calendar&ust=1649078059106132&usg=AOvVaw26wAevExeEfW_AdZSA8UhF
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with David on LinkedIn: https://www.linkedin.com/in/david-bayliss-426556a/
Timestamps:
[00:00] Introduction to David Bayliss
[01:03] Takeaways
[04:56] LexisNexis and David's role
[07:15] Evolution of LexisNexis in 20 years with so many use cases
[08:51] Role of David in structuring data for working with data change
[14:32] Data management and data access
[17:45] Unique challenges of scale, use case, and diversity at LexisNexis
[24:47] Tardis Iron Box
[30:05] Iron Box translation
[32:56] JVM for data science
[34:24] Iron Box meaning
[36:52] Metadata with PII
[39:08] Detrimental privacy / Hairy Kneecap Theory
[40:57] Speeding things up and Anonymized linking
[46:47] What kept David working at LexisNexis?
[50:30] Wrap up
MLOps Coffee Sessions #109 with Ketan Umare, Co-founder and CEO of Union.ai, Why You Need More Than Airflow co-hosted by George Pearse.
// Abstract
Airflow is a beloved tool by data engineers and Machine Learning Engineers alike. But when doing ML what are the shortcomings and why is an orchestration tool like that not always the best developer experience? In this episode, we break down what some key drivers are for using an ML-specific orchestration tool.
// Bio
Ketan Umare is the CEO and co-founder at Union.ai. Previously he had multiple Senior roles at Lyft, Oracle, and Amazon ranging from Cloud, Distributed storage, Mapping (map-making), and machine-learning systems. He is passionate about building software that makes engineers' lives easier and provides simplified access to large-scale systems. Besides software, he is a proud father, and husband, and enjoys traveling and outdoor activities.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Zero to One: Notes on Startups, or How to Build the Future Hardcover by Peter Thiel and Blake Masters:
https://www.amazon.com/Zero-One-Notes-Startups-Future/dp/0804139296
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with George on LinkedIn: https://www.linkedin.com/in/george-pearse-b7a76a157/?originalSubdomain=uk
Connect with Ketan on LinkedIn: https://www.linkedin.com/in/ketanumare/
MLOps Coffee Sessions #108 with Byron Allen, AI & ML Practice Lead at Contino, ML Flow vs Kubeflow 2022 co-hosted by George Pearse.
// Abstract
The amazing Byron Allen talks to us about why MLflow and Kubeflow are not playing the same game!
ML flow vs Kubeflow is more like comparing apples to oranges or as he likes to make the analogy they are both cheese but one is an all-rounder and the other a high-class delicacy. This can be quite deceiving when analyzing the two. We do a deep dive into the functionalities of both and the pros/cons they have to offer.
// Bio
Byron wears several hats. AI & ML practice lead, solutions architect, ML engineer, data engineer, data scientist, Google Cloud Authorized Trainer, and scrum master. He has a track record of successfully advising on and delivering data science platforms and projects. Byron has a mix of technical capability, business acumen, and communication skills that make me an effective leader, team player, and technology advocate.
See Byron write at https://medium.com/@byron.allen
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with George on LinkedIn: https://www.linkedin.com/in/george-pearse-b7a76a157/?originalSubdomain=uk
Connect with Byron on LinkedIn: https://www.linkedin.com/in/byronaallen/
Timestamps:
[00:00] Introduction to Byron Allen
[01:10] Introduction to the new co-host George Pearse
[01:41] ML Flow vs Kubeflow
[05:40] George's take on ML Flow and Kubeflow
[07:28] Writing in YAML
[09:47] Developer experience
[13:38] Changes in ML Flow and Kubeflow
[17:58] Messing around ML Flow Serving
[20:00] A taste of Kubeflow through K-Serve
[23:18] Managed service of Kubeflow
[25:15] How George used Kubeflow
[27:45] Getting the Managed Service
[31:30] Getting Authentication
[32:41] ML Flow docs vs Kubeflow docs
[36:59] Kubeflow community incentives
[42:25] MLOps Search term
[42:52] Organizational problem
[43:50] Final thoughts on ML Flow and Kubeflow
[49:19] Bonus [49:35] Entity-Centric Modeling
[52:11] Semantic Layer options
[57:27] Semantic Layer with Machine Learning
[58:40] Satellite Infra Images demo
[1:00:49] Motivation to move away from SQL
[1:03:00] Managing SQL
[1:05:24] Wrap up
MLOps Coffee Sessions #107 with Ryan Russon, Manager, MLOps and Data Science of Maven Wave Partners, Why and When to Use Kubeflow for MLOps co-hosted by Mihail Eric.
// Abstract
Kubeflow is an excellent platform if your team is already leveraging Kubernetes and allows for a truly collaborative experience.
Let’s take a deep dive into the pros and cons of using Kubeflow in your MLOps.
// Bio
From serving as an officer in the US Navy to Consulting for some of America's largest corporations, Ryan has found his passion in the enablement of Data Science workloads for companies and teams.
Having spent years as a data scientist, Ryan understands the types of challenges that DS teams face in scaling, tracking, and efficiently running their workloads.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://www.mavenwave.com/
https://go.mlops.community/hFApDb
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Ryan on LinkedIn: https://www.linkedin.com/in/ryanrusson/
Timestamps:
[00:00] Introduction to Ryan Russon
[01:13] Takeaways
[04:17] Bullish on KubeFlow!
[06:23] KubeFlow in ML tooling
[11:47] Kubeflow having its velocity
[14:16] To Kubeflow or not to Kubeflow
[18:25] KubeFlow ecosystem maturity
[20:51] Alternatively starting from scratch?
[23:11] Argo workflow vs KubeFlow pipelines
[25:08] KubeFlow as an end-state for citizen data scientists
[28:24] End-to-end workflow key players
[31:17] K-serve
[33:41] KubeFlow on orchestrators
[36:24] Natural transition to KubeFlow maturity
[41:33] "Don't forget about the engineer cost."
[42:21] KubeFlow to other "Flow brothers" trade-offs
[46:12] Biggest MLOps challenge
[49:52] Best practices around file structure
[52:15] KubeFlow changes over the years and what to expect moving forward
[55:52] Best-of-breed vision
[57:54] Wrap up
MLOps Coffee Sessions #106 with Delina Ivanova, Associate Director, Data of HelloFresh, Building a Culture of Experimentation to Speed Up Data-Driven Value co-hosted by Vishnu Rachakonda.
// Abstract
Supply chain/manufacturing are prime areas where the use of data science/analytics/ ML is underdeveloped, and experimentation is required to collect data and enable data-driven solutions.
This talk encourages companies to conduct experiments and collect data over time in order to build accurate/scalable data-driven solutions.
// Bio
Delina has over 10 years of experience across data and analytics, consulting, and strategy with roles spanning financial services, public sector, and CPG industries. She is currently the Associate Director, Data & Insights at HelloFresh Canada where she leads a full-service data team, including data engineering, data science, and business intelligence and automation. She is also a Data Science and Machine Learning instructor in the professional development programs at the University of Toronto and the University of Waterloo.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
The Discourses of Epictetus book: https://www.amazon.com/Discourses-Epictetus/dp/1537427180
The Pyramid Principle: Logic in Writing and Thinking book by Barbara Minto:
https://www.amazon.com/Pyramid-Principle-Logic-Writing-Thinking/dp/0273710516
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Delina on LinkedIn: https://www.linkedin.com/in/delina-ivanova/
Timestamps:
[00:00] Introduction to Delina Ivanova
[00:35] Takeaways
[03:46] Looking for People to organize local Meetups!
[04:30] Delina's career trajectories and growth to the corporate schema
[10:02] Telling stories with data
[13:23] Tricks for being a translator from the business side to data teams
[15:32] Technical engineering management and Delina's day-to-day role
[20:40] Giving up day-to-day individual contributing work and coding
[23:33] Good leadership for technical work
[31:05] Growing team growing productivity
[32:55] Pressured to grow
[35:23] HelloFresh
[39:39] Challenges of e-commerce, CPG, Logistics, and grocery combined
[41:08] Cultural differences
[46:04] Rapid fire session
[52:20] Wrap up
MLOps Coffee Sessions #106 with Curtis Northcutt, CEO & Co-Founder of Cleanlab, Cleanlab: Labeled Datasets that Correct Themselves Automatically co-hosted by Vishnu Rachakonda.
// Abstract
Pioneered at MIT by 3 Ph.D. Co-Founders, Cleanlab is an open-source/SaaS company building the premier data-centric AI tools workflows for (1) automatically correcting messy data and labels, (2) auto-tracking of dataset quality over time, (3) automatically finding classes to merge and delete, (4) auto ml for data tasks, (5) obtaining and ranking high-quality annotations, and (6) training ML models with messy data.
Most of the prescriptive tasks (finding issues) can be done in one line of code with their open-source product: https://github.com/cleanlab/cleanlab.
// Bio
Curtis Northcutt is the CEO and Co-Founder of Cleanlab focused on making AI work reliably for people and their messy, real-world data by automatically fixing issues in any ML dataset. Curtis completed his Ph.D. in Computer Science at MIT, receiving the MIT Thesis Award, NSF Fellowship, and the Goldwater Scholarship. Prior to Cleanlab, Curtis worked at AI research groups including Google, Oculus, Amazon, Facebook, Microsoft, and NASA.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
https://github.com/cleanlab/cleanlab
https://cleanlab.ai/blog/cleanlab-history/
https://labelerrors.com/ https://l7.curtisnorthcutt.com/
https://nips.cc/Conferences/2021/ScheduleMultitrack?event=47102
https://www.youtube.com/watch?v=ieUOv1sQPlw
https://cleanlab.typeform.com/to/NLnU1XZF
Cameo cheating detection system: https://arxiv.org/ftp/arxiv/papers/1508/1508.05699.pdf
The Cathedral & the Bazaar book: https://www.amazon.com/Cathedral-Bazaar-Musings-Accidental-Revolutionary/dp/0596001088
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Curtis on LinkedIn: https://www.linkedin.com/in/cgnorthcutt/
Timestamps:
[00:00] Introduction to Curtis Northcutt
[00:30] Difference between MLOps and Data-Centric AI
[04:04] Realizing the problem of data quality in ML manifesting
[05:11] Computer vision problems
[06:54] War story that got Curtis into Data-Centric AI
[13:50] Overview of Curtis' vision
[14:45] PU Learning
[21:25] Consistency Rate and Flipping Rate
[25:25] One line of code
[29:48] Models makes mistakes
[33:09] Cleanlab play with the environment
[36:30] How ML Engineers should approach data quality problem
[42:42] Quantum computing
[46:39] Result of confident learning
[52:31] Utility for small data sets
[53:53] Cleanlab's huge success stories
[56:13] Rapid fire questions
[58:58] Cloudy and mystified space
[1:03:46] Cleanlab is hiring!
[1:05:06] Wrap up
MLOps Coffee Sessions #104 with the creator of Apache Airflow and Apache Superset Maxime Beauchemin, Future of BI co-hosted by Vishnu Rachakonda.
// Abstract
// Bio
Maxime Beauchemin is the founder and CEO of Preset. Original creator of Apache Superset. Max has worked at the leading edge of data and analytics his entire career, helping shape the discipline in influential roles at data-dependent companies like Yahoo!, Lyft, Airbnb, Facebook, and Ubisoft.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://www.rungalileo.io/
Trade-Off: Why Some Things Catch On, and Others book by Kevin Maney:
https://www.amazon.com/Trade-Off-Some-Things-Catch-Others/dp/0385525958
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Max on LinkedIn: https://www.linkedin.com/in/maximebeauchemin/
Timestamps:
[00:00] Introduction to Maxime Beauchemin
[01:28] Takeaways
[03:42] Paradigm of data warehouse
[06:38] Entity-centric data modeling
[11:33] Metadata for metadata
[14:24] Problem of data organization for a rapidly scaling organization
[18:36] Machine Learning tooling as a subset or of its own
[22:28] Airflow: The unsung hero of the data scientists
[27:15] Analyzing Airflow
[30:44] Disrupting the field
[34:45] Solutions to the ladder problem of empowering exploratory work and mortals superpowers with data
[38:04] What to watch out for when building for data scientists
[41:47] Rapid fire questions
[51:12] Wrap up
MLOps Coffee Sessions #103 with Corey Zumar, MLOps Podcast on Making MLflow co-hosted by Mihail Eric.
// Abstract
Because MLOps is a broad ecosystem of rapidly evolving tools and techniques, it creates several requirements and challenges for platform developers:
- To serve the needs of many practitioners and organizations, it's important for MLOps platforms to support a variety of tools in the ecosystem. This necessitates extra scrutiny when designing APIs, as well as rigorous testing strategies to ensure compatibility.
- Extensibility to new tools and frameworks is a must, but it's important not to sacrifice maintainability. MLflow Plugins (https://www.mlflow.org/docs/latest/plugins.html) is a great example of striking this balance.
- Open source is a great space for MLOps platforms to flourish. MLflow's growth has been heavily aided by: 1. meaningful feedback from a community of ML practitioners with a wide range of use cases and workflows & 2. collaboration with industry experts from a variety of organizations to co-develop APIs that are becoming standards in the MLOps space.
// Bio
Corey Zumar is a software engineer at Databricks, where he’s spent the last four years working on machine learning infrastructure and APIs for the machine learning lifecycle, including model management and production deployment. Corey is an active developer of MLflow. He holds a master’s degree in computer science from UC Berkeley.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Corey on LinkedIn: https://www.linkedin.com/in/corey-zumar/
Timestamps:
[00:00] Origin story of MLFlow
[02:12] Spark as a big player
[03:12] Key insights
[04:42] Core abstractions and principles on MLFlow's success
[07:08] Product development with open-source
[09:29] Fine line between competing principles
[11:53] Shameless way to pursue collaboration
[12:24] Right go-to-market open-source
[16:27] Vanity metrics
[18:57] First gate of MLOps drug
[22:11] Project fundamentals
[24:29] Through the pillars
[26:14] Best in breed or one tool to rule them all
[29:16] MLOps space mature with the MLOps tool
[30:49] Ultimate vision for MLFlow
[33:56] Alignment of end-users and business values
[38:11] Adding a project abstraction separate from the current ML project
[42:03] Implementing bigger bets in certain directions
[44:54] Log in features to experiment page
[45:46] Challenge when operationalizing MLFlow in their stack
[48:34] What would you work on if it weren't MLFlow?
[49:52] Something to put on top of MLFlow
[51:42] Proxy metric
[52:39] Feature Stores and MLFlow
[54:33] Lightning round [57:36] Wrap up
MLOps Coffee Sessions #102 with Yash Sheth, Fixing Your ML Data Blindspots co-hosted by Adam Sroka.
// Abstract
Improving your dataset quality is absolutely critical for effective ML. Finding errors in your datasets is generally a slow, iterative, and painstaking process.
Data scientists should be proactively fixing their model’s blindspots by improving their training data. In this talk, Yash discusses how Galileo helps data scientists identify, fix, and track data across the entire ML workflow.
// Bio
Co-founder and VP of Engineering. Prior to starting Galileo, Yash spent the last decade working on Automatic Speech Recognition (ASR) at Google, leading their core speech recognition platform team, that powers speech-to-text across 20+ products at Google in over 80 languages along with thousands of businesses through their Cloud Speech API.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://www.rungalileo.io/
Trade-Off: Why Some Things Catch On, and Others book by Kevin Maney:
https://www.amazon.com/Trade-Off-Some-Things-Catch-Others/dp/0385525958
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Yash on LinkedIn: https://www.linkedin.com/in/yash-sheth-72111216/
Timestamps:
[00:00] Introduction to Yash Sheth
[02:53] Takeaways
[04:35] Why unstructured data?
[06:59] Fitting in the workflow
[10:56] Digging into the different pains
[18:23] Vision around the democratization of machine learning
[24:31] Unstructured data problem
[25:49] Galileo handling unified tools
[27:21] Calculus for ML
[28:45] Gatekeep
[29:49] Synthetic data in the unstructured data world of Galileo
[33:10] Tips for data scientists that have unstructured data but with a small data set
[35:00] Benefits of users from Galileo
[37:15] Business case for dummies
[42:36] War stories
[44:49] Rapid fire questions
[50:55] Wrap up
MLOps Coffee Sessions #101 with Piero Molino, Declarative Machine Learning Systems: Big Tech Level ML Without a Big Tech Team co-hosted by Vishnu Rachakonda.
// Abstract
Declarative Machine Learning Systems are the next step in the evolution of Machine Learning infrastructure.
With such systems, organizations can marry the flexibility of low-level APIs with the simplicity of AutoML.
Companies adopting such systems can increase the speed of machine learning development, reaching the quality and scalability that only big tech companies could achieve until now, without the need for a team of several thousand people.
Predibase is the turnkey solution for adopting declarative ML systems at an enterprise scale.
// Bio
Piero Molino is CEO and co-founder of Predibase, a company redefining ML tooling. Most recently, he has been Staff Research Scientist at Stanford University working on Machine Learning systems and algorithms in Prof. Chris Ré's' Hazy group. Piero completed a Ph.D. in Question Answering at the University of Bari, Italy. Founded QuestionCube, a startup that built a framework for semantic search and QA. Worked for Yahoo Labs in Barcelona on learning to rank, IBM Watson in New York on natural language processing with deep learning, and then joined Geometric Intelligence, where he worked on grounded language understanding.
After Uber acquired Geometric Intelligence, Piero became one of the founding members of Uber AI Labs. At Uber, he worked on research topics including Dialogue Systems, Language Generation, Graph Representation Learning, Computer Vision, Reinforcement Learning, and Meta-Learning. He also worked on several deployed systems like COTA, an ML and NLP model for Customer Support, Dialogue Systems for driver's hands-free dispatch, the Uber Eats Recommender System with graph learning and collusion detection. He is the author of Ludwig, a Linux-Foundation-backed open source declarative deep learning framework.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: http://w4nderlu.st
http://ludwig.ai https://medium.com/ludwig-ai
Declarative Machine Learning Systems paper By Piero Molino, Christopher Ré: https://cacm.acm.org/magazines/2022/1/257445-declarative-machine-learning-systems/fulltext
Slip of the Keyboard by Sir Terry Pratchett: https://www.terrypratchettbooks.com/books/a-slip-of-the-keyboard/
The Listening Society book series by Hanzi Freinacht: https://www.amazon.com/Listening-Society-Metamodern-Politics-Guides-ebook/dp/B074MKQ4LR
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Piero on LinkedIn: https://www.linkedin.com/in/pieromolino/?locale=en_US
Lightning Sessions #1 with Peeyush Agarwal, Scaling Real-time Machine Learning at Chime.
// Abstract
In this Lighting Talk, Peeyush Agarwal explains 2 key pieces of the ML infrastructure at Chime. Peeyush goes into detail about the current feature store design and feature monitoring process along with the ML monitoring setup.
This Lighting Talk is brought to you by arize.com reach out to them for all of your ML monitoring needs.
// Bio
Peeyush Agarwal is the Lead Software Engineer, ML Platform at Chime. He leads the team which enables data science all the way from exploration, model development, and training to orchestrating batch and real-time models in shadow and production. Earlier, Peeyush was a founding engineer in Chime's DSML team and worked on both building models and getting them into production.
Before Chime, Peeyush was a software engineer at Google where he developed unsupervised ML models that run on Google's data across search, Chrome, YouTube, and other properties to identify intent and use it for personalized ads and recommendations. At Google, he also worked on ML-powered Adaptive Brightness and Adaptive Battery which were launched into Android. Prior to joining Google, Peeyush was an entrepreneur who founded a customer engagement platform that counted Aurelia, Reebok, W, and Red Chief among its clients.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
arize.com
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Peeyush on LinkedIn: https://www.linkedin.com/in/apeeyush/
Timestamps:
[00:00] Introduction to Peeyush Agarwal
[01:08] Agenda
[01:27] What Chime is and what Chime do
[01:44] Chime's products
[02:27] Data Science and Machine Learning at Chime
[08:06] Chime's first real-time model
[08:09] Preventing fraud on Pay Friends
[11:01] Feature Store: Unblock real-time capability
[12:40] Preventing fraud on Pay Friends: Monitoring
[13:35] Preventing fraud on Pay Friends: Instrumentation
[14:36] Monitoring: 4 diverse ways to triage
[15:27] Examples of Metrics: Feature and Model Metrics
[16:39] Scaling Real-time ML at Chime
[17:09] Scaling Real-time ML: Monitoring and Alerting
[18:28] Scaling Real-time ML: Build tools
[20:13] Scaling Real-time ML: Infrastructure Orchestration
[21:36] Scaling Real-time ML: Lessons
MLOps Coffee Sessions #100 with Matthijs Brouns, MLOps Critiques co-hosted by David Aponte.
// Abstract
MLOps is too tool-driven, don't let FOMO drive you to pick the latest feature/model/evaluation/ store but pay closer attention to what you actually need to release more safely and reliably.
// Bio
Matthijs is a Machine Learning Engineer, active in Amsterdam, The Netherlands. His current work involves training MLEs at Xccelerated.io. This means Matthijs divides his time between building new training materials and exercises, giving live trainings, and acting as a sparring partner for the Xccelerators at their partner firms, as well as doing some consulting work on the side.
Matthijs spent a fair amount of time contributing to their open scientific computing ecosystem through various means. He maintains open source packages (scikit-lego, seers) as well as co-chairs the PyData Amsterdam conference and meetup.
// MLOps
Jobs board https://mlops.pallet.xyz/jobs
// Related Links
https://www.youtube.com/watch?v=appLxcMLT9Y
https://www.youtube.com/watch?v=Z1Al4I4Os_A
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Matthijs on LinkedIn: https://www.linkedin.com/in/mbrouns/
Timestamps:
[00:00] Introduction to Matthijs Brouns
[00:28] Takeaways
[03:09] Best of Slack Newsletter
[03:38] AI MLFlow
[04:43] Nanny ML
[05:08] Best confinement buy over the last 2 years
[06:35] Matthijs' day-to-day
[08:24] What's hot right now?
[09:36] ML space, orchestration, deployment
[10:21] Scaling
[13:20] Low-risk releases
[15:27] Scale Limitations or Fundamental in API
[16:33] MLOps maturity to a certain point
[18:57] Interdisciplinary leverage need
[21:11] PyScript
[22:41] Next pipeline tools
[24:02] General pattern to build your own tools
[30:25] Technology recommendation to a chaotic space
[33:46] Structured data vs tabular data
[35:52] Big barriers in production
[37:57] Standardization
[39:20] Automation tension between the engineering side and data science side
[41:50] Low-hanging fruit
[42:30] Human check
[43:43] Rapid fire questions
[48:30] PyData Meetups
MLOps Coffee Sessions #99 with Ronen Dar and Gijsbert Janssen van Doorn, Getting the Most Out of your AI Infrastructure co-hosted by Vishnu Rachakonda.
// Abstract
Run:AI is building a cloud-based platform for building with AI. In this talk, we hear all about why this need exists, how this works, and what value it creates.
// Bio
Ronen Dar
Run:AI Co-founder and CTO Ronen was previously a research scientist at Bell Labs and has worked at Apple and Intel in multiple R&D roles. As CTO, Ronen manages research and product roadmap for Run:AI, a startup he co-founded in 2018. Ronen is the co-author of many patents in the fields of storage, coding, and compression. Ronen received his B.S., M.S., and Ph.D. degrees from Tel Aviv University.
Gijsbert Janssen van Doorn
Gijsbert is Director of Technical Product Marketing at Run:AI. He is a passionate advocate for technology that will shape the future of how organizations run AI. Gijsbert comes from a technical engineering background, with six years in multiple roles at Zerto, a Cloud Data Management and Protection vendor.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
The Hard Thing About Hard Things: Building a Business When There Are No Easy Answers by Ben Horowitz ebook: https://www.scribd.com/book/211302755/The-Hard-Thing-About-Hard-Things-Building-a-Business-When-There-Are-No-Easy-Answers?utm_medium=cpc&utm_source=google_search&utm_campaign=3Q_Google_DSA_NB_RoW&utm_term=&utm_device=c&gclid=Cj0KCQjw1ZeUBhDyARIsAOzAqQLnUzXlgFT1PjU_M6jGqRZmwLbcK-mbfKQI4XrZJBRwgUs4x5j2hQ4aAmt1EALw_wcB
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Ronen on LinkedIn: https://www.linkedin.com/in/ronen-dar/
Connect with Gijsbert on LinkedIn: https://www.linkedin.com/in/gijsbertjvd/
Timestamps:
[00:00] Introduction to Ronen Dar & Gijsbert Janssen van Doorn
[01:25] Takeaways
[04:24] Thank you Run:AI for sponsoring this episode!
[05:13] Run:AI products and components
[09:27] Companies coming to Run:AI and problems they solve
[13:30] Why is this problem hard?
[18:56] Run:AI's Vision
[22:12] Run-on-the-mill workload
[25:36] Engineering challenges and requirements building Run:AI
[32:47] Process of solving problems on the same page
[35:45] Power to give data scientists
[37:38] Avoiding horror stories that might cost a lot of money
[44:23] Running multiple models on a single GPU
[47:17] Never scale down to zero
[48:28] So many ML Start-ups in Israel
[53:00] Vision for the future at GPUs and how will Kubernetes advance
[55:55] Future of AI accelerators
[57:03] Lightning round
[1:02:26] Wrap up
MLOps Coffee Sessions #98 with Brannon Dorsey, Racing the Playhead: Real-time Model Inference in a Video Streaming Environment co-hosted by Vishnu Rachakonda.
// Abstract
Runway ML is doing an incredibly cool workaround applying machine learning to video editing. Brannon is a software engineer there and he’s here to tell us all about machine learning in video and how Runway maintains their machine learning infrastructure.
// Bio
Brannon Dorsey is an early employee at Runway, where he leads the Backend team. His team keeps infrastructure and high-performance models running at scale and helps to enable a quick iteration cycle between the research and product teams.
Before joining Runway, Brannon worked on the Security Team at Linode. Brannon is also a practicing artist who uses software to explore ideas of digital literacy, agency, and complex systems.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: https://brannon.online
Blog: https://runwayml.com/blog/distributing-work-adventures-queuing-and-autoscaling/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Brannon on LinkedIn: https://www.linkedin.com/in/brannon-dorsey-79b0498a/
Timestamps:
[00:00] Introduction to Brannon Dorsey
[00:56] Takeaways
[05:42] Runway ML
[07:00] Replacement for Imovie?
[09:07] Machine Learning use cases of Runway ML
[10:40] Journey of starting as a model zoo to video editor
[14:42] Rotoscoping
[16:23] Intensity of ML models in Runway ML and engineering challenges
[19:55] Deriving requirements
[23:10] Runway's model perspective
[25:25] Why browser hosting?
[27:19] Abstracting away hardware
[32:04] Kubernetes is your friend
[35:29] Statelessness is your friend
[38:17] Merge to master quickly
[42:57] Brannon's winding history becoming an engineer
[46:49] How much do you use Runway?
[49:37] Last book read
[50:36] Last bug smashed
[52:21] MLOps marketing that made eyes rolling
[54:11] Bullish on technology that might surprise people
[54:39] Spot by netapp
[56:42] Implementing Spot by netapp
[56:55] How do you want to be remembered?
[57:22] Wrap up
MLOps Coffee Sessions #97 with Jacob Tsafatinos, Real-Time Exactly-Once Event Processing with Apache Flink, Kafka, and Pinot co-hosted by Mihail Eric.
// Abstract
A few years ago Uber set out to create an ads platform for the Uber Eats app that relied heavily on three pillars; Speed, Reliability, and Accuracy. Some of the technical challenges they were faced with included exactly-once semantics in real-time. To accomplish this goal, they created the architecture diagram above with lots of love from Flink, Kafka, Hive, and Pinot. You can dig into the whole paper (https://go.mlops.community/k8gzZd) to see all the reasoning for their design decisions.
// Bio
Jacob Tsafatinos is a Staff Software Engineer at Elemy. He led the efforts of the Ad Events Processing system at Uber and has previously worked on a range of problems including data ingestion for search and machine learning recommendation pipelines. In his spare time, he can be found playing lead guitar in his band Good Kid.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Uber blog
https://eng.uber.com/author/jacob-tsafatinos/
https://eng.uber.com/real-time-exactly-once-ad-event-processing/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Jacob on LinkedIn: https://www.linkedin.com/in/jacobtsaf/
Timestamps:
[00:00] Introduction to Jacob Tsafatinos
[00:40] Takeaways
[04:25] Jacob's band
[05:29] Lyrics about software engineers or artistic stuff
[06:20] Connection of hobby and real-time system
[08:43] How to game Spotify Algorithm?
[10:00] Data stack for analytics
[13:28] Uber blog
[16:28] Video mess up
[17:04] Considerations and importance of the Uber System
[21:22] Challenges encountered through the Uber System journey
[26:06] Crucial to building the system
[28:13] Not exactly real-time
[30:22] Design decisions main questions
[34:23] Testament to OSS
[36:58] Real-time processing systems for analytical use cases vs Real-time processing systems for predictive use cases
[38:46] Real-time systems necessity
[41:04] Potential that opens up new doors
[41:40] Runaway or learn it?
[46:09] Real-time use case target
[49:31] Resource constrained
[50:48] ML Oops stories
[52:45] Wrap up
MLOps Coffee Sessions #96 with Sebastián Ramírez, FastAPI for Machine Learning co-hosted by Adam Sroka.
// Abstract
Fast API almost never happened. Sebastián Ramírez, the creator of FastAPI, tried as hard as possible not to build something new. After many failed attempts at finding what he was looking for he decided to scratch his own itch and build a new product.
The conversation goes over what Fast API is, how Sebastián built it, what the next big problems to tackle in ML are, and how to focus on adding value where you can.
// Bio
👋 Sebastián Ramírez is the creator of FastAPI, Typer, and other open-source tools.
Currently, Sebastián is a Staff Software Engineer at Forethought while also helping other companies as an external consultant.🤓
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: https://tiangolo.com/
https://fastapi.tiangolo.com/
https://typer.tiangolo.com/
https://www.forethought.ai/
https://sqlmodel.tiangolo.com/
https://github.com/tiangolo
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Sebastián on LinkedIn: https://www.linkedin.com/in/tiangolo/
Timestamps:
[00:00] Introduction to Sebastián Ramírez
[00:44] Takeaways
[02:45] Apply () Conference is coming up!
[03:38] FastAPI background
[05:02] Ramp up reason
[06:17] Tipping point
[08:11] Surprising ways using FastAPI
[10:08] Twist it and break it lessons learned
[12:00] Length of comprehension process
[15:59] Missing pieces
[21:25] Advice to technically capable what to start with
[25:19] Making FastAPI better
[27:52] What to simplify and why are they cumbersome right now?
[30:14] Building FastAPI vs solving the problem
[32:42] Next itch to scratch
[34:26] Landscape's pathway
[38:03] Things that would not change
[40:13] Sebastián's change in life since FastAPI
[43:09] Sebastián's famous tweet
[44:13] Experienced vs inexperienced
[46:07] Approach to becoming a tools expert
[50:22] Wrap up
MLOps Coffee Sessions #95 with Ciro Greco, MLOps as Tool to Shape Team and Culture.
// Abstract
Good MLOps practices are a way to operationalize a more “vertical” practice and blur the boundaries between different stages of “production-ready”. Sometimes you have this idea that production-ready means global availability but with ML products that need to be constantly tested against real-world data, we believe production-ready should be a continuum and that the key person that drives that needs to be the data scientist or the ML engineer.
// Bio
Ciro Greco, VP of AI at Coveo. Ph.D. in Linguistics and Cognitive Neuroscience at Milano-Bicocca. Ciro worked as visiting scholar at MIT and as a post-doctoral fellow at Ghent University.
In 2017, Ciro founded Tooso.ai, a San Francisco-based startup specializing in Information Retrieval and Natural Language Processing. Tooso was acquired by Coveo in 2019. Since then Ciro has been helping Coveo with DataOps and MLOps throughout the turbulent road to IPO.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Company Website
psicologia.unimib.it/03_persone/scheda_personale.php?personId=518
gist.ugent.be/members
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Ciro on LinkedIn: https://www.linkedin.com/in/cirogreco/en
Timestamps:
[00:00] Introduction to Ciro Greco
[02:32] Ciro's bridge to Coveo
[07:15] Coveo in a nutshell
[11:30] Confronting disorganization and challenges
[16:08] Fundamentals of use cases
[18:09] Immutable data in the data warehouse
[21:36] Data management in Coveo
[24:48] Pain for advancement
[29:56] Rational process and Stack
[32:24] Habits of high-performing ML Engineers
[35:46] Sharpening the sword
[37:50] Attracting talents vs firing people
[42:18] Wrap up
MLOps Coffee Sessions #94 with Mark Freeman, Traversing the Data Maturity Spectrum: A Startup Perspective.
// Abstract
A lot of companies talk about having ML and being data-driven, but few are there currently and doing it well. If anything, many companies are on the cusp of implementing ML rather than being ML mature.
As a startup, what decisions are we making today to drive data maturity and set us up for success when we further implement ML in the near future. What business cases are we making for leadership buy-in to invest in data infrastructure as compared to product development while we identify product-market-fit.
// Bio
Mark is a community health advocate turned data scientist interested in the intersection of social impact, business, and technology. His life’s mission is to improve the well-being of as many people as possible through data—especially among those marginalized.
Mark received his M.S. from the Stanford School of Medicine where he was trained in clinical research, experimental design, and statistics with an emphasis on observational studies. In addition, Mark is also certified in Entrepreneurship and Innovation from the Stanford Graduate School of Business.
He is currently a senior data scientist at Humu where he builds data tools that drive behavior change to make work better. His core responsibilities center around 1) building data products that reach Humu's end users, 2) providing product analytics for the product team, and 3) building data infrastructure and driving data maturity.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: humu.com
The Informed Company: How to Build Modern Agile Data Stacks that Drive Winning Insights book:
https://www.amazon.com/Informed-Company-Cloud-Based-Explore-Understand/dp/1119748003
Fundamentals of Data Engineering book by Joe Reis and Matt Housley:
https://www.oreilly.com/library/view/fundamentals-of-data/9781098108298/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Mark on LinkedIn: https://www.linkedin.com/in/mafreeman2/
Timestamps:
[00:00] Introduction to Mark Freeman
[01:43] Grab your own Apron merch @ https://mlops.community/!
[03:41] LinkedIn stardom!
[04:40] Followers or connections?
[05:31] Leveraging essential information platform
[08:56] Investment in time spent on creating and working on a social platform
[12:16] Put yourself out there for people to find you
[16:33] Data maturity is a spectrum that takes time to traverse
[23:43] Maturity of path
[28:43] Fundamentals for data products
[33:05] Foundational data capabilities
[37:32] Value of metrics
[41:48] writing reused code timeframe vs working with stakeholders timeframe
[44:11] Wrap up
[45:14] Look for Meetups near you!
MLOps Coffee Sessions #93 with Krishnaram Kenthapadi, Model Monitoring in Practice: Top Trends co-hosted by Mihail Eric
// Abstract
We first motivate the need for ML model monitoring, as part of a broader AI model governance and responsible AI framework, and provide a roadmap for thinking about model monitoring in practice.
We then present findings and insights on model monitoring in practice based on interviews with various ML practitioners spanning domains such as financial services, healthcare, hiring, online retail, computational advertising, and conversational assistants.
// Bio
Krishnaram Kenthapadi is the Chief Scientist of Fiddler AI, an enterprise startup building a responsible AI and ML monitoring platform. Previously, he was a Principal Scientist at Amazon AWS AI, where he led the fairness, explainability, privacy, and model understanding initiatives in the Amazon AI platform. Prior to joining Amazon, he led similar efforts at the LinkedIn AI team and served as LinkedIn’s representative on Microsoft’s AI and Ethics in Engineering and Research (AETHER) Advisory Board. Previously, he was a Researcher at Microsoft Research Silicon Valley Lab. Krishnaram received his Ph.D. in Computer Science from Stanford University in 2006. He serves regularly on the program committees of KDD, WWW, WSDM, and related conferences, and co-chaired the 2014 ACM Symposium on Computing for Development. His work has been recognized through awards at NAACL, WWW, SODA, CIKM, ICML AutoML workshop, and Microsoft’s AI/ML conference (MLADS). He has published 50+ papers, with 4500+ citations and filed 150+ patents (70 granted). He has presented tutorials on privacy, fairness, explainable AI, and responsible AI at forums such as KDD ’18 ’19, WSDM ’19, WWW ’19 ’20 '21, FAccT ’20 '21, AAAI ’20 '21, and ICML '21.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: https://cs.stanford.edu/people/kngk/
https://sites.google.com/view/ResponsibleAITutorial
https://sites.google.com/view/explainable-ai-tutorial
https://sites.google.com/view/fairness-tutorial
https://sites.google.com/view/privacy-tutorial
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mihail on LinkedIn: https://www.linkedin.com/in/mihaileric/
Connect with Krishnaram on LinkedIn: https://www.linkedin.com/in/krishnaramkenthapadi
Timestamps:
[00:00] Introduction to Krishnaram Kenthapadi
[02:22] Takeaways
[04:55] Thank you Fiddler AI for sponsoring this episode!
[05:15] Struggles in Explainable AI
[06:16] Attacking the problem of difficult models and architectures in Explainability
[08:30] Explainable AI prominence
[09:56] Importance of password manager and actual security
[14:27] Role of Education in Explainable AI systems
[18:52] Highly regulated domains in other sectors
[21:12] First machine learning wins
[23:36] Model monitoring
[25:35] Interests in ML monitoring and Explainability
[27:27] Future of Explainability in the wide range of ML models [29:57] Non-technical stakeholders' voice [33:54] Advice to ML practitioners to address organizational concerns [38:49] Ethically sourced data set [42:15] Crowd-sourced labor [43:35] Recommendations to organizations about their minimal explainable product [46:29] Tension in practice [50:09] Wrap up
MLOps Coffee Sessions #92 with Pete Soderling, Building the World's First Data Engineering Conference.
// Abstract
Keep things centered around community building and what he looks for in teams. Folks that are building their community around their tool, what advice do you have for that? What's worth turning into a company?
// Bio
Pete Soderling is the founder of Data Council and the Data Community Fund. As a former software engineer, repeat founder, and investor in more than 40 data-oriented startups, Pete’s lifetime goal is to help 1,000 engineers start successful companies. Most importantly, Pete is a community builder — from his earliest days of working with the data engineering community starting in 2013, he has witnessed the unique power of specialized networks to bring inspiration, knowledge, and support to technical professionals.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: datacouncil.ai
Youtube channel: https://www.youtube.com/c/DataCouncil
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Pete on LinkedIn: https://www.linkedin.com/in/petesoder/
Timestamps:
[00:00] Introduction to Pete Soderling
[04:35] Top takeaways from the World's First Data Engineering Conference
[05:41] Buzz around the conference
[06:37] Intro to Data Council
[09:20] Pete's mission statement with investing
[11:19] Evaluating gaps in the market and who should solve those
[14:45] One company Peter regrets not investing in
[16:41] Repeating the same mistake
[20:07] Recommendations to engineers to become entrepreneurs
[23:30] Questions to consider before investing
[27:37] Things to do and avoid in open-source projects
[31:03] Something popular you disagree
[35:29] Code as an artifact
[39:16] Hypothetical fundraising
[40:53] Wrap up
MLOps Coffee Sessions #91 with Joseph Haaga, The Shipyard: Lessons Learned While Building an ML Platform / Automating Adherence.
// Abstract
Joseph Haaga and the Interos team walk us through their design decisions in building an internal data platform. Joseph talks about why their use case wasn't a fit for off the self solutions, what their internal tool snitch does, and how they use git as a model registry.
Shipyard blogpost series: https://medium.com/interos-engineering.
// Bio
Joseph leads the ML Platform team at Interos, the operational resilience company. He was introduced to ML Ops while working as a Senior Data Engineer and has spent the past year building a platform for experimentation and serving. He lives in Washington, DC, with his dog Cheese.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: https://joehaaga.xyz
Medium: https://medium.com/interos-engineering
Shipyard blogpost series: https://medium.com/interos-engineering
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Joseph on LinkedIn: https://www.linkedin.com/in/joseph-haaga/
Timestamps:
[00:00] Introduction to Joseph Haaga
[02:07] Please subscribe, follow, like, rate, review our Spotify and Youtube channels
[02:31] New! Best of Slack Weekly Newsletter
[03:03] Interos [04:33] Global supply chain
[05:45] Machine Learning use cases of Interos
[06:17] Forecasting and optimization of routes
[07:14] Build, buy, open-source decision making
[10:06] Experiences with Kubeflow
[11:05] Creating standards and rules when creating the platform
[13:29] Snitches
[14:10] Inter-team discussions when processes fall apart
[16:56] Examples of the development process on the feedback of ML engineers and data scientists
[20:35] Preserving flexibility when introducing new models and formats
[21:37] Organizational structure of Interos
[23:40] Surface area for product
[24:46] Use of Git Ops to manage boarding pass
[28:04] Cultural emphasis
[30:02] Naming conventions
[32:28] Benefit of a clean slate
[33:16] One-size-fits-all choice
[37:34] Wrap up
MLOps Coffee Sessions #90 with Valerio Velardo, Bringing Audio ML Models into Production.
// Abstract
The majority of audio/music tech companies that employ ML still don’t use MLOps regularly. In these companies, you rarely find audio ML pipelines which take care of the whole ML lifecycle in a reliable and scalable manner. Audio ML probably pays the price of being a small sub-discipline of ML. It’s dwarfed by ML applications in image processing and NLP.
In audio ML, novelties tend to travel slowly. However, things are starting to change. A few audio and music tech companies are investing in MLOps. Building MLOps solutions for music presents unique challenges because audio data is significantly different from all other data types.
// Bio
Valerio is MLOps Lead at Utopia Music. He’s also an AI audio consultant who helps companies implement their AI music vision by providing technical, strategy, and talent sourcing services.
Valerio is interested both in the R&D and productization (MLOps) aspects of AI applied to the audio and music domains. He's the host of The Sound of AI, the largest YouTube channel and online community on AI audio with more than 22K subscribers.
Previously, Valerio founded and led Melodrive, a tech startup that developed an AI-powered music engine capable of generating emotion-driven video game music in real-time. Valerio earned a Ph.D. in music AI from the University of Huddersfield (UK).
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Valerio's website
https://valeriovelardo.com/
The Sound of AI YouTube channel:
https://www.youtube.com/channel/UCZPFjMe1uRSirmSpznqvJfQ
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Valerio on LinkedIn: https://www.linkedin.com/in/valeriovelardo/
Timestamps:
[00:00] Introduction to Valerio Velardo
[01:28] Please subscribe and rate us!
[02:40] History of Valerio's love for music
[04:12] Intervention of computer science, AI, and Machine Learning in music
[08:06] Experimenting with Machine Learning
[09:25] Environmental Sound AI
[11:05] AI Music
[15:22] Traditional ML life cycle within music tech companies
[18:02] Representation of data
[22:22] Audio being better served in the market
[30:42] Success metrics
[35:17] Challenges when talking to R&D teams
[38:10] Things need to be battle-hardened before production
[39:09] Education process besides Valerio's Youtube channel
[42:38] Rectifying use cases not related to audio
[45:48] Organizing modular blocks building stacks
[47:59] Open-source tools implementation
[50:28] Wrap up
MLOps Coffee Sessions #89 with Gabriel Straub, A Journey in Scaling AI.
// Abstract
Gabriel talks to us about the difficulties of scaling ML products across an organization. He speaks about differences in profiles of data consumers and data producers, and the challenges of educating engineers so they have greater insights into the effects that their changes to the system may have.
// Bio
Gabriel joined Ocado Technology in 2020 as Chief Data Officer, bringing over 10 years of experience in leading data science teams and helping organizations realize the value of their data. At Ocado Technology his role is to help the organization take advantage of data and machine learning so that we can best serve our retail partners and their customers.
Gabriel is a guest lecturer at London Business School and an Honorary Senior Research Associate at UCL. He has also advised start-ups and VCs on data and machine learning strategies. Before joining Ocado, Gabriel was previously Head of Data Science at the BBC, Data Director at notonthehighstreet.com, and Head of Data Science at Tesco.
Gabriel has a MA in Mathematics from Cambridge and an MBA from London Business School.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
Website: https://www.ocadogroup.com/about-us/ocado-technology
Podcast: https://www.reinfer.io/podcast/ai-pioneers-gabriel-straub-chief-data-scientist-ocado
Blog: https://www.ocadogroup.com/technology/blog
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Gabriel on LinkedIn: https://www.linkedin.com/in/gabriel-s-65081521/
Timestamps:
[00:00] Introduction to Gabriel Straub
[03:14] Best of Slack Newsletter
[04:06] Gabriel's best purchase since the pandemic
[05:37] Ocado's background and Gabriel's role
[07:55] Sliding scale of Ocado
[10:05] Different use cases of Ocado
[12:02] Realizing value with Machine Learning
[13:18] How things need to be computed on the edge
[14:51] Ocado's main day-to-day
[16:17] Being generalizable and when to stop
[19:11] The Golden Path
[21:30] Foundational level of maturity
[24:41] Metrics of success
[27:10] Lifespan of a data
[28:49] Hard lessons learned from producers and consumers
[30:19] Internal assessment
[32:50] Evolution of Ocado
[36:58] Rule-based system
[38:58] Putting data science and/or machine learning value in front of the consumers
[41:55] Going past the constraints
[44:24] What holds people back?
[46:30] Instilling cultural value of doing right and well into the company
[49:42] Being defensive talking about AI
[51:44] Ocado is hiring!
MLOps Coffee Sessions #88 with Javier Andres Mansilla, ML Platform Tradeoffs and Wondering Why to Use Them.
// Abstract
Javier runs ML Platform at Mercado Libre. We’re here with Javier because he’s going to tell us about what the ML platform at Mercado Libre looks like granularly, talk about its purpose, lessons, wins, and future improvements, and share with us some of the most challenging use cases they’ve had to engineer around.
// Bio
During the last 3 years building the internal ML platform for Mercado Libre (NASDAQ MELI), the biggest company in Latam, and the eCommerce & fintech omnipresent solution for the continent.
Seasoned entrepreneur and leader, Javier was co-founder and CTO of Machinalis, a hi-end company building Machine Learning since 2010 (yes, before the breakthrough of neuralnets). When Machinalis got acquired by Mercado Libre, that small team evolved to enable Machine Learning as a capability for a tech giant with more 10k devs, impacting the lives of almost 100 million direct users.
On a daily basis, Javier leads not only the tech and product roadmap of their Machine Learning Platform, but also their users' tracking system, the AB Testing framework, and the open-source office.
Javier loves hanging out with family and friends, python, biking, football, carpentry, and slow-paced holidays in nature!
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Javier on LinkedIn: https://www.linkedin.com/in/javimansilla/
Timestamps:
[00:00] Introduction to Javier Andres Mansilla
[02:18] Refresher to what Mercado Libre is
[06:16] Centralization of Machine Learning platform at Mercado Libre
[11:58] Mercado Libre's working size
[16:15] Hitting the scale
[21:07] Driving ML platform vision and team's business metrics
[28:23] Education process of how to use machine learning on the platform
[36:49] Composition of the team members and finding the right people
[43:05] Stakeholders
[45:32] Decision making
[48:51] Wrap up
[49:52] Bonus from Javier
MLOps Coffee Sessions #87 with Kyle Morris, Don't Listen Unless You Are Going to Do ML in Production.
// Abstract
Companies wanting to leverage ML specializes in model quality (architecture, training method, dataset), but face the same set of undifferentiated work they need to productionize the model. They must find machines to deploy their model on, set it up behind an API, make the inferences fast, cheap, reliable by optimizing hardware, load-balancing, autoscaling, clustering launches per region, queueing long-running tasks... standardizing docs, billing, logging, CI/CD that integrates testing, and more.
Banana.dev's aim is to simplify this process for all. This talk outlines our learnings and the trials and tribulations of ML hosting.
// Bio
Hey all! Kyle did self-driving AI @ Cruise, robotics @ CMU, currently in business @ Harvard. Now he's building banana.dev to accelerate ML! Kyle cares about safely building superhuman AI. Our generation has the chance to build tools that advance society 100x more in our lifetime than in all of history, but it needs to benefit all living things! This requires a lot of technical + social work. Let's go!
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// Related Links
kyle.af
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Kyle on LinkedIn: https://www.linkedin.com/in/kylejohnmorris/
Timestamps:
[00:00] Introduction to Kyle Morris
[02:42] banana.dev
[04:43] banana.dev's vision
[06:22] banana.dev's goal beyond the competition
[07:28] Computer vision optimization
[08:46] Common pitfalls
[11:47] Machine Learning Engineering vs Software Engineering
[13:47] Who do you hire?
[15:12] Disconnect in operationalizing
[18:53] Meeting SLOs if stuff is breaking upstream
[19:48] Is breaking upstream a part of quality?
[21:16] Scenario of what to focus on
[24:02] Advice to people dealing with unrealistic expectations
[28:11] Hard truth
[29:35] MLOps Jobs board - https://mlops.pallet.xyz/jobs
[30:42] Don't Listen Unless You Are Going to Do ML in Production
[33:15] Hurdle in productionizing ML systems
[37:56] Chaos engineering
[42:40] War stories
[45:54] Catalyst on changing the original post on Kyle's blog
[50:11] Wrap up
[51:02] Message banana.dev or Kyle if you have questions regarding production. It's free of charge!
MLOps Coffee Sessions #86 with Julien Bisconti, Building ML/Data Platform on Top of Kubernetes.
// Abstract
When building a platform, a good start would be to define the goals and features of that platform, knowing it will evolve. Kubernetes is established as the de facto standard for scalable platforms but it is not a fully-fledged data platform.
Do ML engineers have to learn and use Kubernetes directly?
They probably shouldn't. So it is up to the data engineering team to provide the tools and abstraction necessary to allow ML engineers to do their work.
The time, effort, and knowledge it takes to build a data platform is already quite an achievement. When it is built, one has to maintain it, monitor it, train people to on-call rotation, implement escalation policies and disaster recovery, optimize for usage and costs, secure it and build a whole ecosystem of tools around it (front-end, CLI, dashboards).
That cost might be too high and time-consuming for some companies to consider building their own ML platform as opposed to cloud offering alternatives. Note that cloud offerings still require some of those points but most of the work is already done.
// Bio
Julien is a software engineer turned Site Reliability Engineer. He is a Google developer expert, certified Data Engineer on Google Cloud and Kubernetes Administrator, mentor for Woman Developer Academy and Google For Startups program. He is working on building and maintaining data/ML platform.
// Related Links
https://portal.superwise.ai/
Crossing the River by Feeling the Stones • Simon Wardley • GOTO 2018: https://www.youtube.com/watch?v=2IW9L1uNMCs
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Julien on LinkedIn: https://www.linkedin.com/in/julienbisconti/
Timestamps:
[00:00] French intro by Julien
[00:32] Introduction to Julien Bisconti
[03:35] Arriving at the non-technical side process of MLOps
[06:06] Envious of people with technological problems
[07:27] People problem bandwidth conversation
[11:04] Atomic decision making
[14:20] Advice to developers either to buy or build in their career potential
[18:23] Jobs board - https://mlops.pallet.xyz/jobs
[21:28] Chaos engineering
[26:33] Role of chaos engineering in building production machine learning systems
[32:59] Core challenge of MLOps
[37:04] Standardization on an industry level
[40:30] Reconciliation of trade-offs using Vertex and Sagemaker
[45:21] Crossing the River by Feeling the Stones talk by Simon Wardley
[47:22] Wrap up
MLOps Coffee Sessions #85 with Emmanuel Ameisen, Continuous Deployment of Critical ML Applications.
// Abstract
Finding an ML model that solves a business problem can feel like winning the lottery, but it can also be a curse. Once a model is embedded at the core of an application and used by real users, the real work begins. That's when you need to make sure that it works for everyone, that it keeps working every day, and that it can improve as time goes on. Just like building a model is all about data work, keeping a model alive and healthy is all about developing operational excellence.
First, you need to monitor your model and its predictions and detect when it is not performing as expected for some types of users. Then, you'll have to devise ways to detect drift, and how quickly your models get stale. Once you know how your model is doing and can detect when it isn't performing, you have to find ways to fix the specific issues you identify. Last but definitely not least, you will now be faced with the task of deploying a new model to replace the old one, without disrupting the day of all the users that depend on it.
A lot of the topics covered are active areas of work around the industry and haven't been formalized yet, but they are crucial to making sure your ML work actually delivers value. While there aren't any textbook answers, there is no shortage of lessons to learn.
// Bio
Emmanuel Ameisen has worked for years as a Data Scientist and ML Engineer. He is currently an ML Engineer at Stripe, where he worked on helping improve model iteration velocity. Previously, he led Insight Data Science's AI program where he oversaw more than a hundred machine learning projects. Before that, he implemented and deployed predictive analytics and machine learning solutions for Local Motion and Zipcar. Emmanuel holds graduate degrees in artificial intelligence, computer engineering, and management from three of France’s top schools.
// Related Links
https://www.amazon.com/Building-Machine-Learning-Powered-Applications/dp/149204511X https://
www.oreilly.com/library/view/building-machine-learning/9781492045106/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Emmanuel on LinkedIn: https://www.linkedin.com/in/ameisen/
Timestamps:
[00:00] Introduction to Emmanuel Ameisen
[03:38] Building Machine Learning Powered Applications book inspiration
[05:19] The writing process
[07:04] Over-engineering NLP
[09:13] CV driven development: intentional or natural
[11:09] Attribute to machine learning team
[14:44] Shortening iteration cycle
[16:41] Advice on how to tackle iteration
[20:00] Failure modes
[21:02] Infrastructure Iteration at Stripe
[27:06] Deployment Steps tests challenges
[29:34] "You develop operational excellence by exercising it." - Emmanuel Ameisen
[33:22] Death of a thousand cuts: Balance of work vs productionization piece balance
[36:15] Reproducibility headaches [40:04] Pipelines as software product
[41:25] Get the book Building Machine Learning Powered Applications: Going from Idea to Product book by Emmanuel Ameisen!
[42:04] Takeaways and wrap up
MLOps Coffee Sessions #84 with Ernest Chan, Lessons from Studying FAANG ML Systems.
// Abstract
Large tech companies invest in ML platforms to accelerate their ML efforts. Become better prepared to solve your own MLOps problems by learning from their technology and design decisions.
Tune in to learn about ML platform components, capabilities, and design considerations.
// Bio
Ernest is a Data Scientist at Duo Security. As part of the core team that built Duo's first ML-powered product, Duo Trust Monitor, he faced many (frustrating) MLOps problems first-hand. That led him to advocate for an ML infrastructure team to make it easier to deliver ML products at Duo. Prior to Duo, Ernest worked at an EdTech company, building data science products for higher-ed. Ernest is passionate about MLOps and using ML for social good.
// Related Links
Lessons on ML Platforms — from Netflix, DoorDash, Spotify, and more: https://ernestklchan.medium.com/lessons-on-ml-platforms-from-netflix-doordash-spotify-and-more-f455400115c7
Paper Highlights-Challenges in Deploying Machine Learning: a Survey of Case Studies https://towardsdatascience.com/paper-highlights-challenges-in-deploying-machine-learning-a-survey-of-case-studies-cafe61cfd04c
Choose boring technologies Slideshare by Dan McKinley: https://www.slideshare.net/danmckinley/choose-boring-technology
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Ernest on LinkedIn: https://www.linkedin.com/in/ernest-chan-68245773/
Timestamps:
[00:00] Introduction to Ernest Chan
[01:07] Takeaways
[02:58] Ernest's Lessons on ML Platforms — from Netflix, DoorDash, Spotify, and more blog post
[05:55] Five components of an ML Platform
[10:09] Limitations highlighted in the blog post
[14:41] Level of maturity or completion observed in company efforts
[16:17] Platform/Architecture admired the most
[17:46] Advice to big tech companies
[22:03] Process of needing an infrastructure and aiming towards having a platform
[24:23] Paper Highlights-Challenges in Deploying Machine Learning: a Survey of Case Studies blog post
[26:24] Takeaways from Paper Highlights-Challenges in Deploying Machine Learning
[30:33] Prioritization
[33:04] Delta Lake
[35:27] Model rollouts and shadow mode
[39:23] Are you an ML Engineer or a Data Scientist?
[40:15] Simple route platform vs flexible platform trade-offs
[41:08] Opinionated and simple vs less opinionated and flexible
[43:22] Choose boring technologies Slideshare by Dan McKinley
[44:36] Wrap up
MLOps Coffee Sessions #83 with Vincent Warmerdam, Better Use cases for Text Embeddings.
// Abstract
Text embeddings are very popular, but there are plenty of reasons to be concerned about their applications. There's algorithmic fairness, compute requirements as well as issues with datasets that they're typically trained on.
In this session, Vincent gives an overview of some of these properties while also talking about an underappreciated use-case for the embeddings: labeling!
// Bio
Vincent D. Warmerdam is a senior data professional who worked as an engineer, researcher, team lead, and educator in the past. He's especially interested in understanding algorithmic systems so that one may prevent failure. As such, he has a preference for simpler solutions that scale, as opposed to the latest and greatest from the hype cycle. He currently works as a Research Advocate at Rasa where he collaborates with the research team to explain and understand conversational systems better.
Outside of Rasa, Vincent is also well known for his open-source projects (scikit-lego, human-learn, doubtlab, and more), collaborations with open source projects like spaCy, his blog over at koaning.io, and his calm code educational project.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Vincent on LinkedIn: https://www.linkedin.com/in/vincentwarmerdam/
MLOps Reading Group meeting on February 11, 2022
Reading Group Session about Feature Stores with Matt Delacour and Mike Moran
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Connect with us on LinkedIn: https://www.linkedin.com/company/mlopscommunity/
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Timestamps:
[00:05] Matt's intro
[00:26] Mike's intro
[01:09] Matt’s talk: Feature store system at Shopify
[01:45] What is Shopify?
[02:05] Shopify Use Case
[02:38] Choosing a solution
[03:19] Managed service vs In-house vs Open-source (Feast)
[06:01] Why did we choose Feast?
[11:25] Implementation Strategy (multi-repo vs mono-repo approaches)
[13:01] Mono-repo approach breakdown
[14:30] Internal SDK
[17:01] Q&A: Does feast satisfy scalability for online inference of Shopify latency requirements?
[19:05] Q&A: Do you rely on Feast to serialize data to the online store?
[20:13] Q&A: Is your mono-repo library a subset of Feast?
[21:18] Q&A: Did you consider using git submodules for a multi-repo?
[23:02] Q&A: Are you storing embeddings with Feast?
[24:30] Q&A: Regarding the mono-repo, which modules are responsible for feature engineering? How do you guarantee that different feature engineering can be used across many DS?
[27:58] Mike’s talk (Feature store at Skyscanner)
[28:08] Kaleidoscope System
[28:25] Background and context of the Feature store
[29:30] Initial state of the feature store
[30:13] How does the marketing team also leverage the feature store
[31:04] Current state of the feature store (marketing & machine learning)
[31:44] SDK approach of creating schemas with dataframes (easy access)
[32:16] Reusability across teams among marketing and DS team
[33:06] GDPR constraints
[33:34] Data updates at the feature store
[36:09] Q&A: When a DS updates a feature, how are you communicating that across teams?
[38:25] Q&A: Are you applying different levels of feature engineering to increase the likelihood of a DS going back to a previous checkpoint of processing?
[40:55] Q&A: In what languages are you implementing the feature store?
[44:28] Q&A: Regarding performance-wise, how do you decide what code remains in Apache Spark vs SQL?
[49:00] Wrap-up
MLOps Community Meetup #93! Two weeks ago, we talked to Chad Sanderson, Trustworthy Data for Machine Learning.
//Abstract
The most common challenge for ML teams operating at scale is data quality.
In this talk, Chad discusses how Convoy invested in a large-scale data quality effort to treat data as an API and provide a data change management surface to enable trustworthy machine learning.
// Bio
Chad Sanderson is the Product Lead for Convoy's Data Platform team, which includes the data warehouse, streaming, BI & visualization, experimentation, machine learning, and data discovery.
Chad has built everything from feature stores, experimentation platforms, metrics layers, streaming platforms, analytics tools, data discovery systems, and workflow development platforms. He’s implemented open source, SaaS products (early and late-stage) and has built cutting-edge technology from the ground up. Chad loves the data space, and if you're interested in chatting about it with him, don't hesitate to reach out.
// Related links
----------- ✌️Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Chad on LinkedIn: https://www.linkedin.com/in/chad-sanderson/
MLOps Coffee Sessions #82 with Donna Schut and Christos Aniftos, Practitioners Guide to MLOps.
// Abstract
The "Practitioners Guide to MLOps" introduced excellent frameworks for how to think about the field. Can we talk about how you've seen the advice in that guide applied to real-world systems? Is there additional advice you'd add to that paper based on what you've seen since its publication and with new tools being introduced?
Your article about selecting the right capabilities has a lot of great advice. It would be fun to walk through a hypothetical company case and talk about how to apply that advice in a real-world setting.
GCP has had a lot of new offerings lately, including Vertex AI. It would be great to talk through what's new and what's coming down the line. Our audience always loves hearing how tool providers like GCP think about the problems customers face and how tools are correspondingly developed.
// Bio
Donna Schut
Donna is a Solutions Manager at Google Cloud, responsible for designing, building, and bringing to market smart analytics and AI solutions globally. She is passionate about pushing the boundaries of our thinking with new technologies and creating solutions that have a positive impact. Previously, she was a Technical Account Manager, overseeing the delivery of large-scale ML projects, and part of the AI Practice, developing tools, processes, and solutions for successful ML adoption. She managed and co-authored Google Cloud’s AI Adoption Framework and Practitioners' Guide to MLOps.
Christos Aniftos
Christos is a machine learning engineer with a focus on the end-to-end ML ecosystem. On a typical day, Christos helps Google customers productionize their ML workloads using Google Cloud products and services with special attention on scalable and maintainable ML environments.
Christos made his ML debut in 2010 while working at DigitalMR, where he led a team of data scientists and developers to build a social media monitoring & analytics tool for the Market Research sector.
// Related links:
Select the Right MLOps Capabilities for Your ML Usecase
https://cloud.google.com/blog/products/ai-machine-learning/select-the-right-mlops-capabilities-for-your-ml-use-case
Practitioner's Guide to MLOps white paper
https://services.google.com/fh/files/misc/practitioners_guide_to_mlops_whitepaper.pdf
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Donna on LinkedIn: https://www.linkedin.com/in/donna-schut/
Connect with Christos on LinkedIn: https://www.linkedin.com/in/aniftos/
Timestamps:
[00:00] Introduction to Donna Schut and Christos Aniftos
[05:52] Inspiration of Practitioner's Guide to MLOps paper
[06:57] Model for working with customers
[08:14] Where are we at MLOps?
[10:20] Process of working with customers
[11:30] Overview of processes and capabilities outlined in Practitioner's Guide to MLOps paper
[16:16] Continuous Training maturity levels
[22:37] Context about the discovery process
[25:21] Disciplines and security mix tend to see
[26:12] Is there a level up in maturity?
[29:50] Success or failures that stand out
[38:00] War stories
[43:16] Internal study of qualities of the best ML engineers
MLOps Coffee Sessions #81 with Davis Treybig and Leigh Marie Braswell, Machine Learning from the Viewpoint of Investors.
// Abstract
Machine learning is a rapidly evolving space that can be hard to keep track of. Every year, thousands of research papers are published in the space, and hundreds of new companies are built both in applied machine learning as well as in machine learning tooling.
In this podcast, we interview two investors who focus heavily on machine learning to get their take on the state of the machine learning industry today: Leigh-Marie Braswell at Founders Fund and Davis Treybig at Innovation Endeavors. We discuss their perspectives on opportunities within MLOps and applied machine learning, common pitfalls and challenges seen in machine learning startups, and new projects they find exciting and interesting in the space.
// Bio
Davis Treybig
Davis (email: [email protected]) is currently a principal on the investment team at Innovation Endeavors, an early-stage venture firm focused on highly technical companies. He primarily focuses on software infrastructure, especially data tooling and security. Prior to Innovation Endeavors, Davis was a product manager at Google, where he worked on the Pixel phone and the developer platform for the Google Assistant. Davis studied computer science and electrical engineering in college.
Leigh Marie Braswell
Leigh Marie (Twitter: @LM_Braswell) is an investor at Founders Fund. Before joining Founders Fund, she was an early engineer & the first product manager at Scale AI, where she originally built & later led product development for the LiDAR/3D annotation products, used by many autonomous vehicles, robots, and AR/VR companies as a core step in their machine learning lifecycles. She also has done software development at Blend, machine learning at Google, and quantitative trading at Jane Street.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Leigh on LinkedIn: https://www.linkedin.com/in/leigh-marie-braswell/
Connect with Davis on LinkedIn: https://www.linkedin.com/in/davistreybig/
MLOps Coffee Sessions #80 with Ale Solano, The Journey from Data Scientist to MLOps Engineer.
// Abstract
After years of failed POCs then all of a sudden one of our models is accepted and will be used in production. The next morning we are part of the main scrum stand-up meeting and a DevOps guy is assisting us. A strange feeling, unknown to us until then, starts growing on the AI team: we are useful!
Deploying models to production is challenging, but MLOps is more than that. MLOps is about making an AI team useful and iterative from the beginning. And it requires a role that takes care of the technical challenges that this implies, given the experimental nature of the ML field, while also serving the product and business needs. If your AI team does not include this role, maybe it's your time to step up and do it yourself! Today, we will chat with Ale about the transition from being a data scientist to a self-called MLOps engineer. And yes, you'll need to study computer science.
// Bio
Ale is born and raised in a mid-small town near Malaga in southern Spain. Ale did his bachelor's degree in robotics because it sounded cool and then he got into machine learning because it was even cooler.
Ale worked in two companies as an ML developer. Now he's on a temporary hiatus to study business and computer science and get a motivation boost.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Ale on LinkedIn: https://www.linkedin.com/in/alesolano/
MLOps Coffee Sessions #79 with Orr Shilon, Platform Thinking: A Lemonade Case Study.
// Abstract
This episode is the epitome of why people listen to our podcast. It’s a complete discussion of the technical, organizational, and cultural challenges of building a high-velocity, machine learning platform that impacts core business outcomes.
Orr tells us about the focus on automation and platform thinking that’s uniquely allowed Lemonade’s engineers to make long-term investments that have paid off in terms of efficiency. He tells us the crazy story of how the entire data science team of 20+ people was supported by only 2 ML engineers at one point, demonstrating the leverage their technical strategy has given engineers.
// Bio
Orr is an ML Engineering Team Lead at Lemonade, currently working an ML Platform, empowering Data Scientists to manage the ML lifecycle from research to development and monitoring.
Previously, Orr worked at Twiggle on semantic search, at Varonis on data governance, and at Intel. He holds a B.Sc. in Computer Science and Psychology from Tel Aviv University.
Orr also enjoys trail running and sometimes races competitively.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Orr on LinkedIn: https://www.linkedin.com/in/orrshilon/
MLOps Coffee Sessions #78 with Erica Greene and Seoyoon Park, Calibration for ML at Etsy - apply() special.
// Abstract
This is a special conversation about Machine Learning calibration at Etsy. Demetrios sat down with Erica Greene and Seoyoon Park to hear about how they implemented Calibration into the Etsy Machine Learning workflow.
The conversation is a pre-chat with these two before their presentation at the apply() conference on February 10th.
Register here: applyconf.com
// Bio
Erica Geen
Erica is an engineering manager with a background in machine learning. She's passionate about developing programs and policies that support women and other underrepresented groups in technology.
Seoyoon Park
Backend software engineer and aspiring software architect interested in producing scalable, performant, and fault-tolerant applications by keeping up to date with best practices and industry standards. Seoyoon strives to better himself and his peers by advocating for frequent knowledge transfers and promoting a culture of continuous learning. Constantly looking for opportunities to grow as a developer and become a leader of the industry.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Erica on LinkedIn: https://www.linkedin.com/in/ericagreene/
Connect with Seoyoon on LinkedIn: https://www.linkedin.com/in/seoyoonpark/
MLOps Coffee Sessions #77 with Scott Hirleman, Data Mesh - The Data Quality Control Mechanism for MLOps?
// Abstract
Scott covers what is a data mesh at a high level for those not familiar. Data mesh is potentially a great win for ML/MLOps as there is very clear guidance on creating useful, clean, well-documented/described and interoperable data for "unexpected use". So instead of data spelunking being a harrowing task, it can be a very fruitful one. And that one data set that was so awesome?
Well, it wasn't a one-off, it's managed as a product with regular refreshes! And there is a LOT more ownership/responsibility on data producers to make sure the downstream doesn't break. Might sound like kumbaya for MLOps (or total BS?) re far cleaner data and fewer upstream breaks so let's discuss the realities and limitations!
// Bio
A self-professed "chaotic (mostly) good character", Scott is focused on helping the data mesh community accelerate towards finding solutions for some of data management's hardest challenges. He founded the Data Mesh Learning community specifically to gather enough people to exchange ideas - much of which is patterned off the MLOps community. He hosts the Data Mesh Radio podcast, where he dives deep into topics related to data mesh to provide the data community with useful perspectives and thoughts on data mesh.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Scott on LinkedIn: https://www.linkedin.com/in/scotthirleman/
MLOps Coffee Sessions #76 with Mohamed Elgendy, Build a Culture of ML Testing and Model Quality.
// Abstract
Machine learning engineers and data scientists spend most of their time testing and validating their models’ performance. But as machine learning products become more integral to our daily lives, the importance of rigorously testing model behavior will only increase.
Current ML evaluation techniques are falling short in their attempts to describe the full picture of model performance. Evaluating ML models by only using global metrics (like accuracy or F1 score) produces a low-resolution picture of a model’s performance and fails to describe the model performance across types of cases, attributes, scenarios.
It is rapidly becoming vital for ML teams to have a full understanding of when and how their models fail and to track these cases across different model versions to be able to identify regression. We’ve seen great results from teams implementing unit and functional testing techniques in their model testing. In this talk, we’ll cover why systematic unit testing is important and how to effectively test ML system behavior.
// Bio
Mohamed is the Co-founder & CEO of Kolena and the author of the book “Deep Learning for Vision Systems”. Previously, he built and managed AI/ML organizations at Amazon, Twilio, Rakuten, and Synapse. Mohamed regularly speaks at AI conferences like Amazon's DevCon, O'Reilly's AI conference, and Google's I/O.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with Mohamed on LinkedIn: https://www.linkedin.com/in/moelgendy/
MLOps Coffee Sessions #75 with Shreya Shankar, Towards Observability for ML Pipelines.
// Abstract
Achieving observability in ML pipelines is a mess right now. We are tracking thousands of means, percentiles, and KL divergences of features and outputs in a haphazard attempt to figure out when and how to retrain models.
In this session, we break down current unsuccessful approaches and discuss the path towards effectively maintaining ML models in production. Along the way, we introduce mltrace -- a preliminary open source project striving towards "bolt-on" observability in ML pipelines.
// Bio
Shreya Shankar is a computer scientist living in the Bay Area. She's interested in building systems to operationalize machine learning workflows. Shreya's research focus is on end-to-end observability for ML systems, particularly in the context of heterogeneous stacks of tools.
Currently, Shreya is doing her Ph.D. in the RISE lab at UC Berkeley. Previously, she was the first ML engineer at Viaduct, did research at Google Brain, and completed her BS and MS in computer science at Stanford University.
// Related Links
Shreya Shankar's blogposts: https://www.shreya-shankar.com/
Shreya Shankar's Podcasts: https://www.listennotes.com/top-episodes/shreya-shankar/
The deployment phase of machine learning by Benedict Evans: https://www.ben-evans.com/benedictevans/2019/10/4/machine-learning-deployment
Shreya Shrankar's mltrace blogpost: https://www.shreya-shankar.com/introducing-mltrace/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Shreya on LinkedIn: https://www.linkedin.com/in/shrshnk
Timestamps:
[00:00] Introduction to Shreya Shankar
[01:12] Shreya's background
[03:22] Contrast in scale influence
[05:28] Embedding ML and building machine learning infused products
[07:26] Management structure and professional incentive
[08:25] Organizational side of MLOps retros
[10:15] Tooling implementations
[12:00] Structured rational investment hardships
[13:17] Working at a start-up
[14:02] Academic work and entrepreneurial ambitions
[16:00] ML Monitoring Observability interest
[17:14] Where to get started
[20:47] Realization while at Viaduct
[23:30] Preventing alert fatigue
[27:04] Tooling bridging the gap
[30:40] Juncture at overall MLOps ecosystem
[33:58] The deployment phase of machine learning - it's the new SQL by Benedict Evans
[35:30] Model monitoring
[36:16] mltrace
[38:28] Introducing mltrace blog post series
[41:25] Tips to our content creators/writers
[43:47] Monitoring through the lens of the database
[47:37] Advice about picking up ML engineering and ML systems development in 2022
[49:36] Database low down the stack
[50:51] Most excited about 2022
[52:13] What MLOps space/ecosystem should change?
[53:21] Funding has changed the incentives around innovation
[54:52] Competition in million-dollar rounds
[55:25] Starting a company
[56:30] Wrap up
MLOps Coffee Sessions #74 with Jesse Johnson, Scaling Biotech.
// Abstract
Scaling a biotech research platform requires managing organization complexity - teams, functions, projects - rather than just the traditional volume, velocity, and variety. By examining the processes and experiments that drive the platform, you can focus your work where it matters the most by finding the ideal balance for each type of experiment along with a number of common trade-offs.
// Bio
Jesse Johnson is head of Data Science and Data Engineering at Dewpoint Therapeutics, an R&D-stage biotech startup. His interest in exploring complex systems, understanding what makes them tick, then using this understanding to improve and scale them led him from academic mathematics, into software engineering (Google, Verily Life Sciences), and then to Biotech (Sanofi, Cellarity, Dewpoint). His goal is to identify ways to scale biotech research through better software and organizational design.
// Related Links
Jessie's blogposts: scalingbiotech.com
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Jesse on LinkedIn: https://www.linkedin.com/in/jesse-johnson-51619a7/
Timestamps:
[00:00] Introduction to Jesse Johnson
[05:10] Jesse's background
[05:52] Biotech environments
[06:31] Jesse's background in Biotech companies
[09:21] Jesse's journey from academic to software engineering
[12:20] Transition from primary output insights/research into writing code
[14:54] Actual hands-on use case in practice
[19:19] Jesse's career trajectory
[23:57] Where we're at state-of-the-art data engineering and its outstanding challenges
[26:50] Dewpoint's data and machine learning challenges and tooling
[29:04] Dewpoint's team structure
[30:20] Jesse being the VP of Data Science and Data Engineering
[33:24] New biotech data makes it hard to design a data platform
[35:35] Changes in how biotech data is viewed
[35:54] Experiment data output
[40:19] Solving challenges in structuring real-world context into interpretable data fields
[44:16] Maturity between the current data engineering and MLOps tooling space
[47:31] Achieving a blogpost mission in 2022
[49:50] Wrap up
MLOps Coffee Sessions #73 with Breno Costa and Matheus Frata, On Structuring an ML Platform 1 Pizza Team.
// Abstract
Breno and Matheus were part of an organizational change at Neoway in recent years. With the creation of cross-functional and platform teams in order to improve the value stream generated by these. They share their experience in creating a machine learning platform team. The challenges they faced along the way, how they approached using product thinking and the results achieved so far.
// Bio
Matheus Frata Matheus is an Electronics Engineer that got into Data Science by accident! During his graduation, Matheus joined Neoway as a Data Scientist, but during that time he saw a lot of problems that were related to engineers! This was Matheus' beginning with MLOPS. Today, Matheus works as a Machine Learning Engineer helping their Data Scientists to FLY!!!
Breno Costa
Breno uses his mixed background in Computer Science and Mathematical Modeling to design and develop ML-based software products. A brief period as an entrepreneur gives a different look at how to approach problems and generate more value. He has worked at Neoway for three years and currently works as a machine learning engineer on the Platform team.
// Related links
https://mlops.community/building-neoways-ml-platform-with-a-team-first-approach-and-product-thinking/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Breno on LinkedIn: https://www.linkedin.com/in/breno-c-costa/
Connect with Matheus on LinkedIn: https://www.linkedin.com/in/matheus-frata/
Timestamps:
[00:00] Introduction to Breno Costa & Matheus Frata
[02:08] Breno's background in Neoway
[03:23] What does Neoway do and Matheus' background in Neoway
[05:43] Organizational structure of Neoway
[07:31] Concept of redesign
[10:47] Getting the structure right as a priority
[15:26] Designing the teams
[20:28] Three different ways of setting up the cells interaction
[23:58] Platform differences
[25:33] Technical components before redesigning and organizational overhauling
[31:50] Supporting platform teams
[33:23] Settling tech stack managing technical needs
[42:10] Building internal tools
[50:10] Wrap up
MLOps Coffee Sessions #72 with Vishnu Rachakonda and Demetrios Brinkmann, 2021 MLOps Year in Review.
// Abstract
Vishnu and Demetrios sit down to reflect on some of the biggest news and learnings from 2021 from the biggest funding rounds to best insights. The two finish out the chat by talking about what to expect in 2022.
// Bio
Demetrios Brinkmann
At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter.
Vishnu Rachakonda
Vishnu is the operations lead for the MLOps Community and co-hosts the MLOps Coffee Sessions podcast. He is a machine learning engineer at Tesseract Health, a 4Catalyzer company focused on retinal imaging. In this role, he builds machine learning models for clinical workflow augmentation and diagnostics in on-device and cloud use cases. Since studying bioengineering at Penn, Vishnu has been actively working in the fields of computational biomedicine and MLOps. In his spare time, Vishnu enjoys suspending all logic to watch Indian action movies, playing chess, and writing.
//Related links
Dr. Angela Duckworth's book on Grit featuring Cody Coleman:
https://www.scribd.com/book/311311935/Grit?utm_medium=cpc&utm_source=google_search&utm_campaign=3Q_Google_DSA_NB_RoW&utm_device=c&gclid=CjwKCAjw0a-SBhBkEiwApljU0klle1jhwhK1hrCtdOzR2NIqNu1Y1D9kkGhFg5k2jvo5cCft7UOCqBoCsigQAvD_BwE
You don't need Kafka Vicki Boykis blog:
https://vicki.substack.com/p/you-dont-need-kafka?s=r
The Informed Company book:
https://www.amazon.com/Informed-Company-Cloud-Based-Explore-Understand/dp/1119748003
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Timestamps:
[01:03] Campfire
[01:31] What are you most interested in learning about?
[02:00] Learning about serving models
[03:42] 2021 MLOps Community growth
[04:22] Engaging people coming back to the community
[05:41] Consistently high-quality interactions
[07:07] Vishnu's 2021 favorite moment in the Coffee Sessions
[10:05] Dr. Angela Duckworth's book on Grit featuring Cody Coleman
[11:43] Biggest surprise over the year for Demetrios
[13:48] You don't need Kafka Vicki Boykis blog
[16:26] What excites Vishnu in 2022
[18:04] The Informed Company book
[20:48] What excites Demetrios in 2022
[26:28] News and blurbs
[33:25] Spinouts
[34:30] Last year's cool events
[36:02] Community progress
[38:47] Community highlights
[41:28] New projects
[44:26] A controversial blogpost
[46:03] Milestones
[46:57] Lessons
[50:00] Shout out and thanks to our sponsors!
Loblaws is one of Canada’s largest grocery store chains, Mefta's team at Loblaw Digital runs several ML systems such as search, recommendations, inventory, and labor prediction on production. In this conversation, he shares his experience setting up their ML platform on GCP using Vertex AI and open-source tools.
The goal of this platform is to help all the data science teams within their organization to take ML projects from EDA to production rapidly while ensuring end-to-end tracking of these ML pipelines. We also talk about our overall platform architecture and how the MLOps tools fit into the end-to-end ML pipeline.
//Bio
Mefta Sadat is a Senior ML Engineer at Loblaw Digital. He has been here for over three years building the Data Engineering and Machine Learning platform. He focuses on productionizing ML services, tools, and data pipelines. Previously Mefta worked at a Toronto-based Video Streaming Company and designed and built the recommendation system for the Zoneify App from scratch. He received his MSc in Computer Science from Ryerson University focusing on research to mitigate risk in Software Engineering using ML.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Timestamps:
[00:00] Introduction to Mefta Sadat
[01:04] Mefta's background
[02:45] Mefta's journey in ML Engineering
[04:19] Use cases of Machine Learning at Loblaws
[06:00] Loblaws' team operation
[07:37] Number of people in the team and number of users in the platform
[08:40] Software engineering process
[10:47] Data platform vs ML platform
[13:10] Timeline leveraging machine learning in Loblaws products and business
[15:01] Transition from legacy systems to the cloud
[16:47] Recommendation System use case - Legacy Style Stack and its impact on the business
[21:01] Biggest challenges and pain points
[24:31] Choices of tools to use
[27:31] Dealing with data access
[30:39] The good, the bad, and the ugly
[32:48] Setting up alerts on image classification models
[33:53] Productionizing ML passion
[36:00] Post-deployment monitoring of recommendation systems
[37:47] Wrap up
MLOps Coffee Sessions #70 with Reah Miyara, 2022 Predictions for MLOps and the Industry.
// Abstract MLOps has moved fast in the last year. What will 2022 be like in the MLOps ecosystem? Raeh from Arize AI comes on to talk to us about what he expects for the new year.
Arize is kindly offering 20 free subscriptions to their tool. No marketing BS these are design partners. First come first serve https://arize.com/mlops-signup/!
// Bio
Reah Miyara is a Senior Product Manager at Arize AI, a leading ML monitoring and observability platform counted on by top enterprises to track billions of predictions daily. Reah joins Arize from Google AI, where he led product strategy for the Algorithms and Optimization organization. His experience as a team and product leader is extensive, touching a broad cross-section of the AI technology landscape.
Reah played pivotal roles in ML and AI initiatives at Google, IBM Watson, Intuit, and NASA Jet Propulsion Laboratory and his work have directly contributed to many important innovations and successes that have moved the broader industry forward. Reah also co-led the Google Research Responsible AI initiative, confronting the risks of AI being misused and taking steps to minimize
AI’s negative influence on the world.
// Relevant Links
Subscription - https://arize.com/mlops-signup/
https://arize.com/blog/welcome-to-arize-reah/
https://arize.com/blog/best-practices-in-ml-observability-for-monitoring-mitigating-and-preventing-fraud/
https://www.reah.me/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletter and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Reah on LinkedIn: https://www.linkedin.com/in/reah/
Timestamps:
[00:00] Introduction to Reah Miyara
[01:57] Wrong predictions
[03:41] Real predictions for 2022
[04:00] One: AI fairness and bias issues will get worse before they get better.
[07:27] Two: Enterprises will stop shipping AI blinds
[10:51] Three: The Citizen Data Scientist will rise
[17:07] Four: The ML infrastructure ecosystem will get more complex
[22:28] Five: Unleash the power of unstructured data
[26:34] Six: Robustness of ML Models against changes
[33:18] We want to have the best ML monitoring and observability tool out there.
[34:07] Demetrios' prediction: More talks about laws and regulations will happen but nothing will actually get done.
[35:27] Wrap up
MLOps Coffee Sessions #69 with James Lamb, Building for Small Data Science Teams co-hosted by Adam Sroka.
// Abstract
In this conversation, James shares some hard-won lessons on how to effectively use technology to create applications powered by machine learning models.
James also talks about how making the "right" architecture decisions is as much about org structure and hiring plans as it is about technological features.
// Bio
James Lamb is a machine learning engineer at SpotHero, a Chicago-based parking marketplace company. He is a maintainer of LightGBM, a popular machine learning framework from Microsoft Research, and has made many contributions to other open-source data science projects, including XGBoost and prefect. Prior to joining SpotHero, he worked on a managed Dask + Jupyter + Prefect service at Saturn Cloud and as an Industrial IoT Data Scientist at AWS and Uptake. Outside of work, he enjoys going to hip hop shows, watching the Celtics / Red Sox, and watching reality TV (he wouldn’t object to being called “Bravo Trash”).
// Relevant Links
James keeps track of conference and meetup talks he has given at https://github.com/jameslamb/talks#gallery. The audience for this podcast might be most interested in "Scaling LightGBM with Python and Dask" and "How Distributed LightGBM on Dask Works".
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Connect with James on LinkedIn: https://www.linkedin.com/in/jameslamb1/
Timestamps:
[00:00] Introduction to James Lamb
[01:11] James' background in the machine learning space
[03:24] LightGBM
[09:56] Community behind LightGBM
[13:36] Background of James in SpotHero
[20:06] Experience in Maturity Models
[22:40] Bottlenecks of tradeoffs between speed and confidence
[28:28] Tools to be excited about
[31:46] To code your own that's already out there
[36:33] Building design decisions
[39:36] Risk of the unicorn
[42:44] Cross team empathy
[47:18] Proudest technical accomplishment and/or biggest frustration less proud of lessons learned
[50:53] SpotHero is hiring!
[51:20] Wrap up
[51:53] Please like, subscribe, and you can leave a review!
MLOps Coffee Sessions #68 with Chris Albon, Wikimedia MLOps co-hosted by Neal Lathia.
// Abstract
// Bio
Chris spent over a decade applying statistical learning, artificial intelligence, and software engineering to political, social, and humanitarian efforts. He is the Director of Machine Learning at the Wikimedia Foundation. Previously, Chris was the Director of Data Science at Devoted Health, Director of Data Science at the Kenyan startup BRCK, cofounded the AI startup Yonder, created the data science podcast Partially Derivative, was the Director of Data Science at the humanitarian non-profit Ushahidi, and was the director of the low-resource technology governance project at FrontlineSMS. Chris also wrote Machine Learning For Python Cookbook (O’Reilly 2018) and created Machine Learning Flashcards.
Chris earned a Ph.D. in Political Science from the University of California, Davis researching the quantitative impact of civil wars on health care systems. He earned a B.A. from the University of Miami, where he triple majored in political science, international studies, and religious studies.
// Relevant Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Neal on LinkedIn: https://www.linkedin.com/in/nlathia/
Connect with Chris on LinkedIn: https://www.linkedin.com/in/chrisralbon/
Timestamps:
[00:00] Introduction to Chris Albon
[00:28] Do you sleep? :-)
[02:43] ML at Wikimedia
[09:27] Wikimedia workflow
[15:00] Creating a repeatable process
[19:11] Wikimedia element team size
[20:47] Wikimedia workflow and hardware
[23:56] Evaluating open source
[29:20] Lacking in ML source tooling
[33:11] Wikimedia's separate data platform
[38:14] Abstractions
[41:50] Experimentation aspect of getting models into production
[44:05] Stack of Abstraction in ML
[47:16] Chris' proudest model
[49:10] How Wikimedia work with communities
[55:24] Large language models
[1:02:16] Beautiful vision
[1:03:23] Wrap up
MLOps Coffee Sessions #67 with John Crousse, ML Stepping Stones: Challenges & Opportunities for Companies co-hosted by Adam Sroka.
// Abstract
In this coffee session, John shares his observations after working with multiple companies which were in the process of scaling up their ML capabilities.
John's observations are mostly around changes in practices, successes, failures, and bottlenecks identified when building ML products and teams from scratch. John shares a few thoughts on building long-term products vs short-term projects, on the important non-ML components, and the most common missing pieces he sees in today's ecosystem. John also elaborates on how those challenges and solutions can differ for different company sizes.
// Bio
John always liked CS/ML/AI but wasn't such a hot topic back then. He found opportunities to work on models in the Financial industry as a consultant from 2007 to 2017 then he went freelance to move outside of the financial industry, and focus on AI/ML.
John likes to do things efficiently, and MLOps is the bottleneck, so he ended up spending more time on MLOPs than models lately.
John finished his CS degree in 2007.
// Relevant Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka
Connect with John on LinkedIn: https://www.linkedin.com/in/john-crousse-31219b9
Timestamps:
[00:00] Introduction to John Crousse
[01:11] Main trends in Machine Learning
[03:07] Symptoms of Machine Learning product
[05:05] Proper product with limited resources
[08:52] Going into production mindsets
[11:22] Bottlenecks and challenges
[14:55] Business case for Machine Learning or MLOps in small organizations
[17:04] Gathering feedbacks best suited to product owners
[19:14] More substantial role
[20:11] Data factory
[24:03] Delivery patterns or tech stacks
[26:06] Bottleneck metrics
[27:28] Concept of evaluation store
[32:18] The biggest gap to bridge
[34:42] Hindrance to people's development
[35:23] "The last mile of the machine learning projects"
[36:40] MLOps assessment survey
[40:10] Who owns the product and path to recommend
[41:34] Datamesh community
[44:41] Tips on balancing between pure autonomy
[45:58] Wrap up
MLOps Coffee Sessions #66 with Jacopo Tagliabue, Machine Learning at Reasonable Scale.
// Abstract
We believe that immature data pipelines are preventing a large portion of industry practitioners from leveraging the latest research on ML: truth is, outside of Big Tech and advanced startups, ML systems are still far from producing the promised ROI.
The good news is that times are changing: thanks to a growing ecosystem of tools and shared best practices, even small teams can be incredibly productive at a “reasonable scale”. Based on our experience as founders and researchers, we present our philosophy for modern, no-nonsense data pipelines, highlighting the advantages of a "PaaS-like" approach.
// Bio
Educated in several acronyms across the globe (UNISR, SFI, MIT), Jacopo Tagliabue was co-founder and CTO of Tooso, an A.I. company in San Francisco acquired by Coveo in 2019. Jacopo is currently the Director of AI at Coveo, shipping models to hundreds of customers and millions of users. When not busy building products, he is exploring topics at the intersection of language, reasoning, and learning: his research and industry work is often featured in the general press and premier A.I. venues. In previous lives, he managed to get a Ph.D., do sciency things for a pro basketball team, and simulate a pre-Columbian civilization.
// Relevant Links
Bigger boat repo: https://github.com/jacopotagliabue/you-dont-need-a-bigger-boat
TDS series: https://towardsdatascience.com/tagged/mlops-without-much-ops (ep 3 and a NEW open-source contribution on data ingestion coming up)
Open datasets for e-commerce and MLops experiments: https://github.com/coveooss/SIGIR-ecom-data-challenge
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Jacopo on LinkedIn: https://www.linkedin.com/in/jacopotagliabue/
MLOps Coffee Sessions #65 with Skylar Payne, The Future of Data Science Platforms is Accessibility.
// Abstract
The machine learning and data science space is blowing up -- new tools are popping up every day. While we seem to have every type of "Flow" and "Store" you could imagine, few people really understand how to glue this stuff together. Despite all the tools we have available, we still see companies failing to leverage data science effectively to drive business results.
Instead of spending time driving business results, data scientists spend their time fiddling with Kubernetes, trying to debug that Spark serialization error figuring out how to map their code into the awkward "AI Pipeline" SDK. We have an industry filled with tools built by engineers... for engineers, rather than for data scientists. It's deeply disempowering.
Meanwhile, data is still used effectively to drive decisions in many companies. Analysts have been solving very similar problems on the back of applications like Excel, Tableau, and Mode for literally decades. While there are still challenges in analytics, the MLOps space could learn something from analytics tools. Analytics tools better understand how to make their tools accessible. Analytics tools better understand the value of iterability. Analytics tools better understand that data problems are wicked problems:
- we have to iterate on the formulation and solution simultaneously
- they involve many stakeholders with different opinions
- there's no "right" answer
- the problems are never 100% solved.
If we're going to really drive the most business value from data science, we need to understand how to design our teams and tools to effectively work against such problems.
The future of data science platforms is accessibility and iterability.
// Bio
Data is a superpower, and Skylar has been passionate about applying it to solve important problems across society. For several years, Skylar worked on large-scale, personalized search and recommendation at LinkedIn -- leading teams to make step-function improvements in our machine learning systems to help people find the best-fit role. Since then, he shifted my focus to applying machine learning to mental health care to ensure the best access and quality for all. To decompress from his workaholism, Skylar loves lifting weights, writing music, and hanging out at the beach!
// Relevant Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
MLOps Reading Group meeting on November 20, 2021
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Connect with us on LinkedIn: https://www.linkedin.com/company/mlopscommunity/
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
MLOps Coffee Sessions #64 with Slater Victoroff, The Future of AI and ML in Process Automation.
// Abstract
The Unstructured Imperative
Recent advances in AI have dramatically advanced the state of the art around unstructured data, especially in the spaces of NLP and computer vision. Despite this, the adoption of unstructured technologies has remained low. Why do you think that is? How have the dynamics changed in the last five years?
Multimodal AI
Historic AI approaches have generally been constrained to one data modality (i.e. text or image). Recently, a wide range of papers in image captioning and document understanding have emphasized the need for more sophisticated "multimodal" techniques which can fuse information from multiple modalities. What is multimodal learning, and why is it so promising? Why are we seeing such an explosion of activity? What is Indico doing in this space?
Machine Teaching
As methods of supervision become more complex and multi-faceted, many researchers have begun investigating the inverse problem. That is how do we design supervision systems that more naturally follow human processes? What are some interesting trends in "the space", and where can we expect this field to go in the next few years?
// Bio
Slater Victoroff is the Founder and CTO of Indico, an enterprise AI solution for unstructured content that emphasizes document understanding.
Slater has been building machine learning solutions for startups, governments, and Fortune 100 companies for the past seven years and is a frequent speaker at AI conferences.
Indico’s framework requires 1000x less data than traditional machine learning techniques, and they regularly beat the likes of AWS, Google, Microsoft, and IBM in head-to-head bake-offs.
// Relevant Links
https://indico.io
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Slater on LinkedIn: https://www.linkedin.com/in/slatervictoroff
Dmytro Dzhulgakov, PyTorch: Bridging AI Research and Production.
Talking PyTorch is always interesting, as the Facebook ML OSS project is one of the most important parts of the machine learning tooling ecosystem. This week, we talked to Dmytro Dzhulgakov, a tech lead for PyTorch.
We started off talking about Dmytro's journey to being an engineer and tech lead at Facebook, and what his role entails. Dmytro has been at Facebook for 10+ years, so he gave some very interesting advice on how to manage a career in software engineering for the machine learning world. After that, we got deep into the present and future of PyTorch and what improvements the project is making to support MLOps workflows. PyTorch is a large project, and Dmytro shared with us the valuable lessons he learned from confronting multifaceted scaling challenges while working on PyTorch. Finally, we talked about the future of machine learning engineering, especially as relates to how software engineers work by comparison.
// Abstract
Over the past few years, PyTorch became the tool of choice for many AI developers ranging from academia to industry. With the fast evolution of state-of-the-art in many AI domains, the key desired property of the software toolchain is to enable the swift transition of the latest research advances to practical applications.
In this coffee session, Dmytro discusses some of the design principles that contributed to this popularity, how PyTorch navigates inherent tension between research and production requirements, and how AI developers can leverage PyTorch and PyTorch ecosystem projects for bringing AI models to their domain.
// Bio
Dmytro Dzhulgakov is a technical lead of PyTorch at Facebook where he focuses on the framework core development and building the toolchain for bringing AI from research to production.
Previously he was one of the creators of ONNX, a joint initiative aimed at making AI development more interoperable. Before that Dmytro built several generations of large-scale machine learning infrastructure that powered products like Ads or News Feed.
// Relevant Links
https://pytorch.org/
https://pytorch.org/blog/
https://ai.facebook.com/blog/pytorch-builds-the-future-of-ai-and-machine-learning-at-facebook/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Dmytro on LinkedIn: https://www.linkedin.com/in/dzhulgakov/
MLOps Coffee Sessions #62 with Joel Grus, MLOps from Scratch.
// Abstract
In this talk, Joel Grus of “I don’t like notebooks” fame shares with us his 2021 perspective on notebooks, where he thinks MLOps is now, and what his hot takes in the data space are now.
// Bio
Joel Grus is a Principal Engineer at Capital Group, where he leads a team that builds search, data, and machine learning products for the investment group. He is the author of the bestselling O'Reilly book *Data Science from Scratch*, the not-bestselling self-published book *Ten Essays on Fizz Buzz*, and the controversial JupyterCon talk "I Don't Like Notebooks." He recently moved to Texas after living in Seattle for a very long time.
// Relevant Links
Data Science from Scratch book: https://www.oreilly.com/library/view/data-science-from/9781491901410/
Data Science from Scratch, 2nd Edition book: https://www.oreilly.com/library/view/data-science-from/9781492041122/
Ten Essays on Fizz Buzz: Meditations on Python, mathematics, science, engineering, and design book: https://www.amazon.com/Ten-Essays-Fizz-Buzz-Meditations/dp/0982481829 or https://leanpub.com/fizzbuzz/
I Don't Like Notebooks talk: https://www.youtube.com/watch?v=7jiPeIFXb6U
I Don't Like Notebooks - #JupyterCon 2018 slides: https://docs.google.com/presentation/d/1n2RlMdmv1p25Xy5thJUhkKGvjtV-dkAIsUXP-AL4ffI/edit#slide=id.g362da58057_0_658
Fizz Buzz in Tensorflow: https://joelgrus.com/2016/05/23/fizz-buzz-in-tensorflow/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Joel on LinkedIn: https://www.linkedin.com/in/joelgrus/
Timestamps:
[00:00] Introduction to Joel Grus
[01:32] Joel's background in tech
[07:47] Joel's I Don't Like Notebooks talk on Jupyter Con
[13:42] Better tooling around notebooks
[16:48] Hex
[17:20] Step function evolution
[20:41] Kinds of professionals required in Joel's organization to practice MLOps
[23:08] Evaluation process
[25:51] Sagemaker bring your own algorithm
[27:30] Flexibility of models
[31:55] Hot takes on data science world
[34:19] Current Overall Maturity of MLOps
[37:23] Kinds of problem in NLP and search
[39:52] Finding ways to put structures
[40:50] Probabilistic nature of machine learning systems
[43:10] Data scientists coping up on writing production code
[46:33] Invaluability of code review
[47:22] Common repo structure
[47:57] Reviewing codes
[49:15] Code pals
[50:36] Readability and function
[52:23] Leverage code review
[53:10] Remote work
MLOps Coffee Sessions #60 with Svet Penkov, ML Tests.
// Abstract
How confident do you feel when you deploy a new model? Does improving an ML model feel like a game of "whack-a-mole"? ML is taking over all sorts of industries and yet ML testing tools are virtually non-existent.
Drawing parallels from software engineering and electronic circuit board design to the aviation and semiconductor industries, the need for principled quality assurance (QA) step in the MLOps pipeline is long overdue. Let's talk about why ML testing is hard, what can we do about it and what place should ML QA take in the future?
// Bio
Svet has been building robots ever since he was a kid. At some point, Svet got interested in not just how to build them, but actually how to make them think, and so he did a Ph.D. in AI & Robotics at the University of Edinburgh, UK. Towards the end of Svet's Ph.D., he joined FiveAI as a Research Scientist and led the motion prediction team for 3 years.
Throughout his career, Svet spent endless hours fixing model regressions and fighting with edge cases and so at some point he had enough of it and decided it's time to do something about it. That's how Svet started Efemarai where they are building a platform for testing and improving ML continuously.
// Relevant Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Svet on LinkedIn: https://www.linkedin.com/in/svpenkov/
Timestamps:
[00:00] Introduction to Svet Penkov
[02:10] Svet's background in tech
[04:34] Testing on robotics vs areas of machine learning
[05:21] What's missing in testing right now?
[08:56] Who should test?
Step 1. Figuring out the requirements
[12:04] Edge cases
Steps 2. Access of variation
[13:29] Step 3. Validation and Verification
[16:15] New challenges that need to be addressed
[18:25] Test-driven development viability argument
[20:26] Software engineering tests vs machine learning engineering tests
[23:23] Rule of tools in MLOps
[26:15] Figuring out the difficulty in designing the API's
[27:48] Svet's vision for the future
[29:15] Moving goal post
[31:00] 10 data points being realistic
[31:27] Getting less
[32:20] Efemarai: Where it came from and Why?
[33:53] Efemarai - Functional Magnetic Resonance Imaging
[35:21] A perfect world journey
[36:22] Value of tests
[37:55] Get ready for the MLOps Community Slack testing channel!
Coffee Sessions #60 with Alexandre Patry, Path to Productivity in Job Search and Job Recommendation AI at LinkedIn.
// Abstract
A year ago, LinkedIn job search and recommendation AI teams were at the end of a growth cycle. They were fighting many fires at once: a high number of user complaints, engineers spending a significant amount of their time keeping our machine learning pipelines running, online infrastructure that wasn't supporting their growth, and challenges ramping new models to experiment. In this talk, Alex discusses how they all came together to manage these challenges and set themselves for their next phase of growth.
// Bio
Alex has been a machine learning engineer at LinkedIn for almost seven years. He had tour of duties in LinkedIn Groups, content search, and discovery, feed, and has been tech leading in LinkedIn Talent Solutions and Careers for the last two years.
Prior to working at LinkedIn, Alex lived in Montreal where he completed a Ph.D. in Statistical Machine Translation, then work for five years on information extraction.
// Relevant Links
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, Feature Store, Machine Learning Monitoring and Blogs: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Skylar on LinkedIn: https://www.linkedin.com/in/skylar-payne-766a1988/
Connect with Alexandre on LinkedIn: https://www.linkedin.com/in/patry/
Coffee Sessions #59 with Cody Coleman, Data Quality Over Quantity or Data Selection for Data-Centric AI.
// Abstract
Big data has been critical to many of the successes in ML, but it brings its own problems. Working with massive datasets is cumbersome and expensive, especially with unstructured data like images, videos, and speech. Careful data selection can mitigate the pains of big data by focusing computational and labeling resources on the most valuable examples.
Cody Coleman, a recent Ph.D. from Stanford University and founding member of MLCommons, joins us to describe how a more data-centric approach that focuses on data quality rather than quantity can lower the AI/ML barrier. Instead of managing clusters of machines and setting up cumbersome labeling pipelines, you can spend more time tackling real problems.
// Bio
Cody Coleman recently finished his Ph.D. in CS at Stanford University, where he was advised by Professors Matei Zaharia and Peter Bailis. His research spans from performance benchmarking of hardware and software systems (i.e., DAWNBench and MLPerf) to computationally efficient methods for active learning and core-set selection. His work has been supported by the NSF GRFP, the Stanford DAWN Project, and the Open Phil AI Fellowship.
// Relevant
Links [preprint] Similarity Search for Efficient Active Learning and Search of Rare Concepts: [https://arxiv.org/abs/2007.00077](https://arxiv.org/abs/2007.00077)
[video] Similarity Search for Efficient Active Learning and Search of Rare Concepts: [https://www.youtube.com/watch?v=vRVyOEK2JUU](https://www.youtube.com/watch?v=vRVyOEK2JUU)
[blog post] Selection via Proxy: Efficient Data Selection for Deep Learning: [https://dawn.cs.stanford.edu/2020/04/23/selection-via-proxy/](https://dawn.cs.stanford.edu/2020/04/23/selection-via-proxy/)
[slides] The DAWN of MLPerf: [https://drive.google.com/file/d/17ZpX0GOtOXG8QMn6KEc_Le8tUfDBlgDE/view](https://drive.google.com/file/d/17ZpX0GOtOXG8QMn6KEc_Le8tUfDBlgDE/view)
[blog post] About Cody's research: [https://hai.stanford.edu/news/cody-coleman-lowering-machine-learnings-barriers-help-people-tackle-real-problems](https://hai.stanford.edu/news/cody-coleman-lowering-machine-learnings-barriers-help-people-tackle-real-problems)
[video] About Cody: [https://www.youtube.com/watch?v=stxJMsxxxtA](https://www.youtube.com/watch?v=stxJMsxxxtA)
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Cody on LinkedIn: https://www.linkedin.com/in/codyaustun/
Timestamps:
[00:00] Introduction to Cody Coleman
[03:10] Cody's life story
[07:35] Cody's journey in tech
[15:04] Interest in Machine Learning and work at Stanford came about
[21:48] Data-centric Machine Learning Data Quality
[28:56] Research and Industry being together
[33:33] Advice to practitioners
[38:03] Principles of Machine Learning in an academic setting
[43:50] Data-centric promising techniques that stand out
[53:51] Developing benchmarks
[56:34] Guardrails for machine learning vs automated testing suites
[1:02:57] Creating something valuable and useful
[1:07:05] Data collecting vs Data Hoarding
Coffee Sessions #58 with Anne Cocos, 10 Types of Features your Location ML Model is Missing.
// Abstract
Machine learning on geographic data is relatively under-studied in comparison to ML on other formats like images or graphs. But geographic data is prevalent across a wide variety of domains (although many practitioners may not think of it that way). Clearly, any dataset with `latitude` and `longitude` columns can be viewed as geographic data, but also any dataset with a `zipcode`, `city`, `address`, or `county` can be construed as geographic. Demographics, weather, foot traffic, points of interest, and topographic features can all be used to enrich a dataset with any of these types of keys.
Incorporating relatively straightforward geographic features into models can yield substantial improvements; adding "distance to the beach" or "square mileage reachable within 10 min drive" to a real estate pricing model, for example, can lead to significant decreases in model error.
Unfortunately, many ML teams find it difficult to incorporate these types of geographic data into their models because the process of ingesting from geographic formats (geojson or shapefiles), projecting, and properly joining with their existing data can be a large infrastructure lift.
In this coffee session, Anne discusses ways to simplify the process of incorporating geographic or location data into the MLOps workflow, as well as interesting trends in the geographic ML research community that will ultimately make it easier for us to learn from geography just as we do with images or graphs today.
// Bio
Dr. Anne Cocos currently leads data science and machine learning at Ask Iggy, Inc., a venture-backed, seed round startup focused on location analytics. Her team builds tools that make it simple for data scientists to leverage location information in their models and analyses. Previously she was the Director and Head, NLP and Knowledge Graph at GlaxoSmithKline, where she built algorithms and infrastructure to enable GSK’s scientists to leverage all the world’s written biomedical knowledge for drug discovery. She also worked on applied natural language processing research at The Children’s Hospital of Philadelphia Department of Biomedical Informatics. Anne completed her Ph.D. in computer science at the University of Pennsylvania, where she was supported by the Google Ph.D. Fellowship and the Allen Institute for Artificial Intelligence Key Scientific Challenges award.
Before shifting her career toward artificial intelligence, Anne spent several years as an end-user of early ML-powered technologies in the U.S. Navy and at HelloWallet. Her previous degrees are from the U.S. Naval Academy, Royal Holloway University of London, and Oxford University. She currently lives just outside Philadelphia with her husband and three boys.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Anne on LinkedIn: https://www.linkedin.com/in/annecocos/
Coffee Sessions #57 with Michael Del Balso and Erik Bernhardsson, The Future of ML and Data Platforms.
// Abstract
Machine learning, data analytics, and software engineering are converging as data-intensive systems become more ubiquitous. Erik Bernhardsson, ex-CTO at Better and former Spotify machine learning lead, and Mike Del Balso, CEO at Tecton and former Uber machine learning lead and co-creator of Michelangelo sit down to chat with us today.
These two jammed with us about building machine learning platform systems and teams, the modern operational data stack and how it allows more machine learning applications to thrive, and how to successfully take advantage of data in the process of building products and companies.
// Bio
Michael Del Balso
Mike is the co-founder of Tecton, where he is focused on building next-generation data infrastructure for Operational ML. Before Tecton, Mike was the PM lead for the Uber Michelangelo ML platform. He was also a product manager at Google where he managed the core ML systems that power Google’s Search Ads business. Previous to that, he worked on Google Maps. He holds a BSc in Electrical and Computer Engineering summa cum laude from the University of Toronto.
Erik Bernhardsson
Erik is currently working on some crazy data stuff since early 2021 but previously spent 6 years as the CTO of Better.com, growing the tech team from 1 to 300. Before Better, Erik spent 6 years at Spotify, building the music recommendation system and managing a team focused on machine learning.
// Relevant Links
Building a Data Team at a Mid-stage Startup: A Short Story
https://erikbern.com/2021/07/07/the-data-team-a-short-story.html
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Mike on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/
Connect with Erik on LinkedIn: https://www.linkedin.com/in/erikbern
Timestamps:
[01:12] Introduction to Michael Del Balso and Erik Bernhardsson
[03:23] High-level space in data
[07:25] Complexity in the data world
[09:13] Data lake + data bricks
[15:20] Platform strategy
[16:05] "Platform is when the economic value of everybody that uses this exceeds the value of the company that creates it." - Bill Gates
[18:17] Centralizing platforms
[21:06] Team spin up centralization or decentralization
[27:18] Manifestations of being too far from a centralized and decentralized platform
[29:24] Centralized vs Decentralized
[33:33] Platform value and appropriate sizing
[35:43] Building a Data Team at a Mid-stage Startup: A Short Story blog post by Erik Bernhardsson
[38:51] Machine Learning as a sub-problem of Data
[42:16] Operational ML
[46:30] Spotify recommendations
[47:13] Real-time data flows at Spotify
[49:40] Data stack, Machine Learning stack, and Back-end stack reusability
[51:40] Container management
Soumanta wouldn't claim they've reached where they want to and they're still learning, so he's happy sharing successes as well as failures at Yugen.ai.
// Abstract
Determining Minimum Achievable Goals helps Yugen.ai ensure a significant amount of focus on value-added and impact before diving deep into solutions & building ML Systems. In this episode, Soumanta discusses Balancing ML Development vs Ops and Monitoring efforts while scaling plus their focus on improvements in small sprints.
Soumanta wouldn't claim they've reached where they want to and they're still learning, so he's happy sharing successes as well as failures at Yugen.ai.
// Bio
Soumanta is a Co-founder at Yugen.ai, an early-stage startup in the Data Science and MLOps space.
We imagine the future to be shaped by the convergence and simultaneous adoption of Algorithms, Engineering and Ops, and Responsible AI. Our mission is to help effectuate and expedite the same for our client partners by creating large-scale, reliable, and personalized ML Systems.
// Relevant Links
A blog Soumanta wrote when Yugen turned one https://medium.com/swlh/yugen-ai-turns-one-1089f3bf169
Presentation, ML REPA 2021 Title of the Talk - Reducing the distance between Prototyping and Production, Why obsessing over experimentation and iteration compounds ROIs
Slides - https://drive.google.com/file/d/1J9Cv6IPPkGpOTq8Xl_AQCKaR0-pKMUmA/view?usp=sharing
Video - https://youtu.be/4PEbgQTw1W0
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Soumanta on LinkedIn: www.linkedin.com/in/soumanta-das/
Timestamps:
[00:00] Introduction to Soumanta Das
[00:24] What's Yugen.ai's name all about?
[02:02] Starting during the pandemic
[05:13] Determination to continue during the pandemic
[08:02] State of the art in Yugen.ai and its future
[11:32] Time to value defining ML to a business
[13:01] Building a strong ML engineering culture
[19:06] Data scientists patterns
[20:00] Helper functions
[22:45] Code review
[25:32] Repeatable use cases
[27:48] Minimum achievable goals
[30:30] Production management goals
[34:30] Use cases and System design document
[36:20] Practices that helped Yugen.ai build ML systems
[40:05] Growing pains in the scaling process
[43:54] Yugen.ai war stories
[46:50] Unrealizing there's something wrong and there's actually something wrong
[48:10] Data observability tools
[49:42] Hands-on deck
Coffee Sessions #55 with Salwa Muhammad, Learning and Teaching MLOps Applications.
//Abstract
Salwa shared her perspective on how FourthBrain and all learners can keep their education strategy fresh enough for the current zeitgeist. Furthermore, Salwa, Demetrios, and Vishnu talked about principles of effective learning that are important to keep in mind while embarking on any educational journey.
This was a great conversation with a lot of practical tips that we hope you all listen to!
// Bio
Salwa Nur Muhammad is the Founder/CEO of FourthBrain, an AI/ML education startup backed by Andrew Ng's AI Fund. FourthBrain trains Machine Learning engineers through a hybrid 2-3 month cohort-based programs that combine accountability of weekly instructor-led live sessions with the flexibility of online content.
Salwa founded FourthBrain after executive leadership roles at Udacity and Trilogy Education Services (acquired by 2U Inc).
She has over 10 years of experience leveraging technology to develop scalable education programs at higher-ed institutions and ed-tech companies, building new business units, launching international programs, and hiring and training cross-functional teams.
// Relevant Links
https://www.fourthbrain.ai/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Salwa on LinkedIn: https://www.linkedin.com/in/salwanur/
Timestamps:
[00:00] Introduction to Salwa Muhammad
[01:20] Salwa's journey in tech
[05:30] Advice to new ML engineers
[10:21] Curriculum development process
[17:36] FourthBrain's current status and what's next
[21:53] Hardest piece in the course
[24:49] Knowing the right job in a role confused world
[30:05] Needing to upskill without going insane
[35:10] Generalist vs Specialist on T-shaped Analogy
[41:15] Counseling learners in terms of long-term progression
[43:00] MLOps trajectories recommendation
Coffee Sessions #54 with Niall Murphy, Machine Learning SRE.
//Abstract
SRE is making its way into the machine learning world. Software engineering for machine learning requires reliability, performance, and maintainability. Site reliability engineering is the field that deals with reliability and ensuring constant, real-time performance. Niall Murphy, most recently Global Head of SRE at Microsoft Azure, helps us understand what SRE can do for modern ML products and teams.
Building machine learning teams requires a diverse set of technical experiences, and Niall shares his thoughts on how to do that most effectively. Machine learning organizations need to start to take advantage of SRE best practices like SLOs, which Niall walks through. Production machine learning depends on high-quality software engineering, and we get Niall's take on how to ensure that in a machine learning context.
// Bio
Niall Murphy has been interested in Internet infrastructure since the mid-1990s. He has worked with all of the major cloud providers from their Dublin, Ireland offices - most recently at Microsoft, where he was global head of Azure Site Reliability Engineering (SRE). His books have sold approximately a quarter of a million copies worldwide, most notably the award-winning Site Reliability Engineering, and he is probably one of the few people in the world to hold degrees in Computer Science, Mathematics, and Poetry Studies. He lives in Dublin, Ireland, with his wife and two children.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Niall on LinkedIn: https://www.linkedin.com/in/niallm/
Timestamps:
[00:00] Introduction to Niall Murphy
[00:36] SRE background to Machine Learning space transition
[07:10] SLO's being a challenge in the ML space
[09:42] SRE Hiring Investments
[15:10] Behavior of teams concept
[17:45] Challenges dealing with ML production
[18:27] Update on Reliable Machine Learning book
[22:46] Monitoring
[25:05] Difference between ML and SRE
[29:18] Incident response in Machine Learning
[34:46] Rollbacks
[35:50] Machine Learning burden overtime
[42:42] Niall's journey to the SRE space and focus to develop himself
Coffee Sessions #53 with David Aponte, Demetrios Brinkmann, and Vishnu Rachakonda, MLOps Insights.
//Abstract
MLOps Insights from MLOps community core organizers Demetrios Brinkmann, Vishnu Rachakonda, and David Aponte. In this conversation the guys do a deep dive on testing with respect to MLOps, talk about what they have learned recently around the ML field, and what new things are happening with the MLOps community.
//Bio
David Aponte
David is one of the organizers of the MLOps Community. He is an engineer, teacher, and lifelong student. He loves to build solutions to tough problems and share his learnings with others. He works out of NYC and loves to hike and box for fun. He enjoys meeting new people so feel free to reach out to him!
Demetrios Brinkmann
At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter.
Vishnu Rachakonda
Vishnu is the operations lead for the MLOps Community and co-hosts the MLOps Coffee Sessions podcast. He is a machine learning engineer at Tesseract Health, a 4Catalyzer company focused on retinal imaging. In this role, he builds machine learning models for clinical workflow augmentation and diagnostics in on-device and cloud use cases. Since studying bioengineering at Penn, Vishnu has been actively working in the fields of computational biomedicine and MLOps. In his spare time, Vishnu enjoys suspending all logic to watch Indian action movies, playing chess, and writing.
Other Links:
Continuous Delivery for Machine Learning article by Martin Fowler: https://martinfowler.com/articles/cd4ml.html
To Engineer Is Human book by Henry Petroski: https://www.amazon.com/Engineer-Human-Failure-Successful-Design/dp/0679734163
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Timestamps:
[00:14] Tests and how to do tests in MLOps
[09:10] Learning from Vishnu and David's new job
[12:42] How will it change?
[19:48] Forcing to do the right thing vs allowing to do the wrong thing
[21:54] Dealing with Machine Learning Models and Data
[25:10] Feature store and monitoring compare page
Coffee Sessions #52 with Dave Bergstein, Vector Similarity Search at Scale.
//Abstract
Ever wonder how Facebook and Spotify now seem to know you better than your friends? Or why the search feature in some products really “gets” you while in other products it feels stuck in the '90s? The difference is vector search— a method of indexing and searching through large volumes of vector embeddings to find more relevant search results and recommendations.
Dave Bergstein, the Director of Product at Pinecone, joins us to describe how vector search is used by companies today, what are the challenges of deploying vector search to production applications, and how teams can overcome those challenges even without the engineering resources of Facebook or Spotify.
// Bio Dave
Bergstein is Director of Product at Pinecone. Dave previously held senior product roles at Tesseract Health and MathWorks where he was deeply involved with productionalizing AI. Dave holds a Ph.D. in Electrical Engineering from Boston University studying photonics. When not helping customers solve their AI challenges, Dave enjoys walking his dog Zeus and CrossFit.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Dave on LinkedIn: https://www.linkedin.com/company/pinecone-io/mycompany/
Coffee Sessions #51 with Sahbi Chaieb, ML security: Why should you care?
//Abstract
Sahbi, a senior data scientist at SAS, joined us to discuss the various security challenges in MLOps. We went deep into the research he found describing various threats as part of a recent paper he wrote. We also discussed tooling options for this problem that is emerging from companies like Microsoft and Google.
// Bio
Sahbi Chaieb is a Senior Data Scientist at SAS, he has been working on designing, implementing, and deploying Machine Learning solutions in various industries for the past 5 years. Sahbi graduated with an Engineering degree from Supélec, France, and holds an MS in Computer Science specialized in Machine Learning from Georgia Tech.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Sahbi on LinkedIn: https://www.linkedin.com/in/sahbichaieb/
Timestamps:
[00:00] Introduction to Sahbi Chaieb
[01:25] Sahbi's background in tech
[02:57] Inspiration of the article
[09:40] Why should you care about keeping our model secure?
[12:53] Model stealing
[14:16] Development practices
[17:24] Other tools in the toolbox covered in the article
[21:29] Stories/occurrences where data was leaked
[24:45] EU Regulations on robustness
[26:49] Dangers of federated learning
[31:50] Tooling status on model security [33:58] AI Red Teams
[36:42] ML Security best practices
[38:26] AI + Cyber Security
[39:26] Synthetic Data
[42:51] Prescription on ML Security in 5-10 years
[46:37] Pain points encountered
Coffee Sessions #50 with Alex Chung and Srivathsan Canchi, Creating MLOps Standards.
// Abstract
With the explosion in tools and opinionated frameworks for machine learning, it's very hard to define standards and best practices for MLOps and ML platforms. Based on their building AWS SageMaker and Intuit's ML Platform respectively, Alex Chung and Srivathsan Canchi talk with Demetrios and Vishnu about their experience navigating "tooling sprawl". They discuss their efforts to solve this problem organizationally with Social Good Technologies and technically with mlctl, the control plane for MLOps.
// Bio
Alex Chung
Alex is a former Senior Product Manager at AWS Sagemaker and an ML Data Strategy and Ops lead at Facebook. He's passionate about the interoperability of MLOps tooling for enterprises as an avenue to accelerate the industry.
Srivathsan Canchi
Srivathsan leads the machine learning platform engineering team at Intuit. The ML platform includes real-time distributed featurization, scoring, and feedback loops. He has a breadth of experience building high scale mission-critical platforms. Srivathsan also has extensive experience with K8s at Intuit and previously at eBay, where his team was responsible for building a PaaS on top of K8s and OpenStack.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Alex on LinkedIn: https://linkedin.com/in/alex-chung-gsd
Connect with Sri on LinkedIn: https://www.linkedin.com/in/srivathsancanchi/
Timestamps:
[00:00] Introduction to Alex Chung and Srivathsan Canchi
[01:36] Alex's background in tech
[03:07] Srivathsan's background in tech
[04:36] What is SGT?
[05:53] 3 Categories of SGT
1. Education
2. Standardization
3. Orchestration
[07:00] Standardization is desirable
[13:03] Perspective from both sides
[13:39] Profile breakdown of Standardization
[17:20] Importance of Standardization in enterprise
[21:02] Tooling sprawl
[24:04] Standardizing the different interfaces between MLOps tools
[31:54] mlctl
[33:35] mlctl's future
[38:38] How mlctl helps the workflow of Intuit
[41:00] CIGS evolve the different spaces
Coffee Sessions #49 with Stefan Krawczyk, Aggressively Helpful Platform Teams.
//Abstract
At Stitch Fix there are 130+ “Full Stack Data Scientists” who in addition to doing data science work, are also expected to engineer and own data pipelines for their production models. One data science team, the Forecasting, Estimation, and Demand team were in a bind. Their data generation process was causing them iteration & operational frustrations in delivering time-series forecasts for the business. the solution? Hamilton, a novel python micro-framework, solved their pain points by changing their working paradigm.
Some of the main workers on Hamilton are the dedicated engineering team called Data Platform. Data Platform builds services, tools, and abstractions to enable DS to operate in a full-stack manner avoiding hand-off. In the beginning, this meant DS built the web apps to serve model predictions, now as the layers of abstractions have been built over time, they still dictate what is deployed, but write much less code.
// Bio
Stefan loves the stimulus of working at the intersection of design, engineering, and data. He grew up in New Zealand, speaks Polish, and spent formative years at Stanford, LinkedIn, Nextdoor & Idibon. Outside of work in pre-covid times Stefan liked to 🏊, 🌮, 🍺, and ✈.
// Other Links
https://www.youtube.com/watch?v=B5Zp_30Knoo
https://www.slideshare.net/StefanKrawczyk/hamilton-a-micro-framework-for-creating-dataframes https://www.slideshare.net/StefanKrawczyk/deployment-for-free-removing-the-need-to-write-model-deployment-code-at-stitch-fix
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Stefan on LinkedIn: https://linkedin.com/in/skrawczyk
Timestamps:
[00:00] Introduction to Stefan Krawczyk
[00:37] Why Hamilton?
[01:50] Stefan's background in tech
[04:15] Model Life Cycle Team
[06:48] Managing outcomes generated by data scientists
[09:04] Teams doing the same thing
[12:41] Vision of getting code down to zero
[18:40] Freedom and autonomy went wrong
[21:17] Sub teams
[24:00] Create and deploy models easily
[24:28] Interesting challenge to define
[25:15] Stitch Fix Model productionization to be proud of
[26:23] Hamilton to open-source
[28:45] Model Envelope
[31:45] Deployment for free
[34:53] Use of Model Envelope in Model Artifact
[37:16] Extending API definition in a model envelope for the model
[39:00] Dependencies [40:08] Monitoring at scale
[43:43] Advice in terms of neat abstraction
[46:19] Envelope vs Container
[47:33] Time frame of Hamilton's development and its benefits
Coffee Sessions #48 with Julien Chaumond, Tour of Upcoming Features on the Hugging Face Model Hub.
//Abstract
Julien Chaumond’s Tour of Upcoming Features on the Hugging Face Model Hub. Our MLOps community guest in this episode is Julien Chaumond the CTO of Hugging Face - every data scientist’s favorite NLP Swiss army knife.
Julien, David, and Demetrios spoke about many topics including:
Infra for hosting models/model hubs
Inference widgets for companies with CPUs & GPUs (for companies)
Auto NLP which trains models
“Infrastructure as a service”
// Bio
Julien Chaumond is Chief Technical Officer at Hugging Face, a Brooklyn and Paris-based startup working on Machine learning and Natural Language Processing, and is passionate about democratizing state-of-the-art AI/ML for everyone.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Julien on LinkedIn: https://www.linkedin.com/in/julienchaumond/
Timestamps:
[00:00] Introduction to Julien Chaumond
[01:57] Julien's background in tech
[04:35] "I have this vision of building a community where the greatest people in AI can come together and basically invent the future of Machine Learning together."
[04:55] What is Hugging Face?
[06:17] "We have the goal of bridging the gap between research and production on actual production use cases."
[06:45] Start of open-source in Hugging Face
[07:50] Chatbox experiment (reference resolution system) - linking pronouns to the subjects of sentences
[10:20] From a project to a company
[11:46] "The goal was to explore in the beginning."
[11:57] Importance of platform
[14:25] "Transfer learning is an efficient way of Machine Learning. Providing your platform around change that people want to start from pre-trained model and fine-tune them into the specific use case is something that can be big so we built some stuff to help people do that."
[15:35] Narrowing down the scope of service to provide
[16:27] "We have some vision of what we want to build but a lot of it is the small incremental improvements that we bring to the platform. I think it's the natural way of building stuff nowadays because Machine Learning is moving so fast."
[20:00] Model Hubs
[22:37] "We're guaranteeing that we don't build anything that introduces any lagging to Hugging Face because we're using Github. You'll have that peace of mind."
[26:31] Storing model artifacts
[27:00] AWS - cache - stored to an edge location all around the globe
[28:39] Inference widgets powering
[27:17] "For each model on the model hub we try to ensure that we have the metadata about the model to be able to actually run it."
[32:11] Deploying infra function
[32:38] "Depending on the model and library, we optimize the custom containers to make sure that they run as fast as possible on the target hardware that we have."
[34:59] "Machine Learning is still pretty much hardware dependent."
[36:11] Hardware usage
[39:04] "CPU is super cheap. If you are able to run Berks served with a 1-millisecond on CPU because you have powerful optimizations, you don't really need GPUs anymore. It's cost-efficient and energy-efficient."
[40:30] Challenges of Hugging Face and what you learned
[41:10] "It may sound like a super cliche but the team that you assembled is everything."
[43:22] War stories in Hugging Face
[44:12] "Our goal is more forward-looking to be helpful as much as we can to the community."
[48:25] Hugging Face accessibility
Coffee Sessions #47 with Jeremy Howard, fast.ai, AutoML, Software Engineering for ML.
//Abstract
Advancement in ML Workflows: You've been around the ML world for long enough to have seen how much workflows, tooling, frameworks, etc. have matured and allowed for greater scale and access. We'd love to reflect on your personal journey in this regard and hear about your early experiences putting models into production, as well as how you appreciate/might improve the process now.
Data Professional Diversity and MLOps: Your work at fast.ai, Kaggle, and now with NBDEV has played a huge part in supercharging a diverse ecosystem of professionals that contribute to ML-like ML/data scientists, researchers, and ML engineers. As the attention turns to putting models into production, how do you think this range of professionals will evolve and work together? How will things around building models change as we build more?
Turning Research into Practice: You've consistently been a leader in applying cutting-edge ideas from academia into practical code others can use. It's one of the things I appreciate most about the fast.ai course and package. How do you go about picking which ideas to invest in? What advice would you give to industry practitioners charged with a similar task at their company?
// Bio
Jeremy Howard is a data scientist, researcher, developer, educator, and entrepreneur. Jeremy is a founding researcher at fast.ai, a research institute dedicated to making deep learning more accessible. He is also a Distinguished Research Scientist at the University of San Francisco, the chair of WAMRI, and is Chief Scientist at platform.ai.
Previously, Jeremy was the founding CEO of Enlitic, which was the first company to apply deep learning to medicine, and was selected as one of the world’s top 50 smartest companies by MIT Tech Review two years running. He was the President and Chief Scientist of the data science platform Kaggle, where he was the top-ranked participant in international machine learning competitions for 2 years running. He was the founding CEO of two successful Australian startups (FastMail, and Optimal Decisions Group–purchased by Lexis-Nexis). Before that, he spent 8 years in management consulting, at McKinsey & Co, and at AT Kearney. Jeremy has invested in, mentored, and advised many startups, and contributed to many open-source projects.
He has many media appearances, including writing for the Guardian, USA Today, and The Washington Post, appearing on ABC (Good Morning America), MSNBC (Joy Reid), CNN, Fox News, BBC, and was a regular guest on Australia’s highest-rated breakfast news program. His talk on TED.com, “The wonderful and terrifying implications of computers that can learn”, has over 2.5 million views. He is a co-founder of the global Masks4All movement.
// Other Links:
jhoward.fastmail.fm
enlitic.com
jphoward.wordpress.com/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Jeremy on LinkedIn: https://www.linkedin.com/in/howardjeremy/
Timestamps:
[00:00] Introduction to Jeremy Howard
[02:11] Jeremy's background
[03:10] Workflow
[12:59] Platform development
[19:53] Balancing API
[22:57] Moment of inefficiency
[27:42] Helpful tactics
[29:05] University of tools evolving
[41:10] Resources to solve problems
[43:30] Jupiter notebooks
[47:20] Putting Jupiter notebooks into production
[48:42] MBDev
[51:20] Jeremy's experiences and frustrations with putting ML into production
Coffee Sessions #46 with Pablo Estevez, What We Learned from 150 Successful ML-enabled Products at Booking.com.
//Abstract
While most of the Machine Learning literature focuses on the algorithmic or mathematical aspects of the field, not much has been published about how Machine Learning can deliver meaningful impact in an industrial environment where commercial gains are paramount. We conducted an analysis on about 150 successful customer-facing applications of Machine Learning, developed by dozens of teams in Booking.com, exposed to hundreds of millions of users worldwide, and validated through rigorous Randomized Controlled Trials. Our main conclusion is that an iterative, hypothesis-driven process, integrated with other disciplines was fundamental to build 150 successful products enabled by Machine Learning.
// Bio
Pablo Estevez is the Principal Data Scientist at Booking.com. He has worked on recommendations, personalization, and experimentation across the Booking.com website, as well as as a manager on several machine learning, data science, and product development teams.
// Other Links
Talk on the topic: https://www.youtube.com/watch?v=ljhtfrtuNqw&t=4h24m30s
The paper: https://blog.kevinhu.me/2021/04/25/25-Paper-Reading-Booking.com-Experiences/bernardi2019.pdf
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Pablo on LinkedIn: https://www.linkedin.com/in/estevezpablo/
MLOps community meetup #70! Last Wednesday, we talked to Monika Venckauskaite, Senior Machine Learning Engineer at Vinted.
//Abstract
One of the areas, that is the most transformed by ML these years is cybersecurity. Traditionally, SIEM (Security Intelligence and Event Management) is performed by human analysts. However, as the cyber powers and tools of the world are growing, we need more and more of these specialists. The entire area of cybersecurity is experiencing a shortage of talent. This is where the ML is coming in to help us. Cybersecurity ML systems require a lot of expertise from specialists as well as unique ways of handling user-sensitive data. This imposes various architectural solutions. In this talk, Monika introduces us to the ways of using ML in cybersecurity and the unique challenges we face.
//Bio
Monika is a keen and curious ML engineer, loving to build systems. She's started in machine learning as a master's student, looking for Higgs Boson and Dark matter within the CERN data. Later on, Monika moved to the IT industry and worked on various machine learning projects, including Open Source Intelligence Tools and a distributed system for ML cybersecurity analytics.
Currently, Monika works as an MLOps engineer, improving the MLOps platform that is used in production to shipping models to a 45 million-user platform. Monika also works in a start-up that is innovating satellite communication. In her free time, she loves books, traveling, and playing some music.
// Takeaways
Cyber threats are all around us. ML as technology is both a savior and a threat.
GDPR and sensitive user data bring in extra challenges for cybersecurity intelligence systems, leading to more complex architectural decisions.
ML helps to fight the talent shortage.
Cybersecurity requires real-time ML systems and reacting ASAP.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Monika on LinkedIn: https://www.linkedin.com/in/monika-in-space/
Timestamps:
[00:00] Introduction to Monika Venčkauskaitė
[05:50] Monika's background in tech
[08:50] Machine Learning in Cyber Security
[09:37] Content
[10:19] Our world is run by machines
[11:16] Cybersecurity Threats
[12:44] Cybersecurity Incident Response
Cycle:
1. Identify
2. Protect
3. Detect
4. Respond
5. Recover
[25:05] The Iceberg
Surface Web - 4% Indexed and easily searchable
Deep Web - 90% Not Indexed, tougher to find
Dark Web - 6% Obscured, difficult to discover
[47:45] Recommendation: AI Superpowers: China, Silicon Valley, And The New World Order by Kai-Fu Lee (https://www.amazon.com/AI-Superpowers-China-Silicon-Valley/dp/132854639X)
[50:54] "I think we are going in the same direction but our implementations are different."
Coffee Sessions #45 with Diego Oppenheimer of Algorithmia, Enterprise Security and Governance MLOps.
//Abstract
MLOps in the enterprise is difficult due to security and compliance. In this MLOps Coffee Session, the CEO of Algorithmia, Diego talks to us about how we can better approach MLOps within the enterprise. This is an introduction to essential principles of security in MLOps and why it is crucial to be aware of security best practices as an ML professional.
// Bio
Diego Oppenheimer is co-founder and CEO of Algorithmia. Previously, he designed, managed, and shipped some of Microsoft’s most used data analysis products including Excel, Power Pivot, SQL Server, and Power BI. He holds a Bachelor’s degree in Information Systems and a Master’s degree in Business Intelligence and Data Analytics from Carnegie Mellon University.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Diego on LinkedIn: https://www.linkedin.com/in/diego/
Timestamps:
[00:00] Thank you Diego and Algorithmia for sponsoring this session!
[01:04] Introduction to Diego Oppenheimer
[02:55] Security
[04:42] "The level of scrutiny for apps and development and that of the operational software is much higher."
[07:40] "We take the Ops part of MLOps very, very seriously and it's really about the operational side of the equation."
[09:22] MLSecOps
[11:42] "The code doesn't change, but things change cause the data changed."
[15:23] Maturity of security
[18:45] "To a certain degree, we have general parameters of software DevOps In software engineering and DevOps, and we're adapting it to this new world of ML."
[19:03] Development workflow
[20:58] "In the ideal world, you're just sitting in your data science platform, your auto ML platform, whatever it is that you're working with, you can push a model."
[22:50] Security, responsibility and authentication
[23:38] "What you don't want to learn is how to do automation every single time there's a new use case. That's just not a good use of your time." [24:30] Hurdles needed to be cleared
[24:47] "I would argue that there's no such thing as Bulletproof in software. That doesn't exist. It never has and never will."
[26:25] Machine Learning security risks
1. Operational risk
2. Brand risk
3. Strategic risk
[28:23] Machine Learning security risk standards
[31:11] "There's a world where you can reverse engineer a model by essentially feeding a whole bunch of data and understanding where that comes back."
[33:55] How to change the mindset of relaxed companies when it comes to security
[35:19] "It takes time and money to figure out security."
[37:52] Conscientious when building systems
[39:44] "Look at the end result of the workflow and understand the value of that workflow, which you should know at that point because if you're going into an ML workflow without understanding what the end value is going to be, it's not a good sign."
[40:19] Root cause analysis
[41:00] Threat modeling
[41:14] "There's a natural next step where there's threat modeling for ML systems and it's a task that gets built and understood, and nobody's going to enjoy doing it."
[43:07] Security as code
[45:29] MLRE
Coffee Sessions #44 with Grant Wright of SEEK Ltd., Autonomy vs. Alignment: Scaling AI Teams to Deliver Value.
/Abstract
Setting AI teams up for success can be difficult, especially when you’re trying to balance the need to provide teams with autonomy to innovate and solve interesting problems while ensuring they are aligned to the organizations' strategy. Operating models, rituals and processes can really help to set teams up for success; but, there is no right answer, and as you scale and priorities change your approach needs to change too.
Grant shares some of his learnings in establishing a cross-functional team of data scientists, engineers, analysts, product managers, and otologists to solve employment information problems at SEEK, and how the team has evolved as they’ve scaled from a team of 30 in Melbourne Australia to over 100 team members across 5 countries in the past three years.
// Bio
Grant heads the Artificial Intelligence & Product Analytics teams at SEEK, where he leads a global team of over 120 Data Scientists, Software Engineers, Ontologists, and AI Product Managers who deliver AI Services to online employment and education platforms across the Asia Pacific and the Americas.
Grant has held various strategy and product and tech leadership roles over the past 15 years, with experience in scaling, AI teams to deliver outcomes across multiple geographies.
Grant holds a Bachelor of Computer and Information Science (Software Development) and a Bachelor of Business (Economics) from the Auckland University of Technology.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Grant on LinkedIn: https://www.linkedin.com/in/wrightgrant/
In this Machine Learning System Design Review, Shaji Chennan Kunnummel walks us through the system design for Pinterest’s near-real-time architecture for detecting similar images. We discuss their usage of Kafka, Flink, rocksdb, and much more. Starting with the high-level requirements for the system, we discussed Pinterest’s focus on debuggability and an easy transition from their batch processing system to stream processing. We then touch on the different system interfaces and components involved such as Manas—Pinterest’s custom search engine—and how it all ends up in their custom graph database, downstream Kafka streams, and to Pinterest’s feature store—Galaxy. With Shaji’s expert knowledge of the system, we were able to do a deep dive into the system’s architecture and some of its components.
// Experiences
15+ years of experience in software product development.
Led multiple teams in a highly agile, collaborative, and cross-functional environment.
Designed and implemented highly scalable, fault-tolerant, and optimized distributed systems that scale to handle millions of requests per second. In-depth knowledge of Object-oriented programming and design patterns in C++/Java/Python/Golang.
Designed and built complex data pipelines and microservices to train and serve machine learning models.
Built analytics pipelines for processing and mining high-volume data set using Hadoop and Map-Reduce frameworks.
In-depth knowledge of distributed storage, consistency models, NoSQL data modeling, Cloud computing environment (AWS and Google Cloud).
MLOps community meetup #69! Last Wednesday we talked to Emmanuel Raj, Senior Machine Learning Engineer at TietoEvry.
//Abstract
The talk focuses on simplifying/demystifying MLOps, encourages others to take steps to learn this powerful SE method. We also talked about Emmanuel's journey in ML engineering, the evolution of MLOps, daily life, and SE problems, and what's next in MLOps (fusion of AIOps, EU AI regulations impact on MLOps workflow, etc).
//Bio
Emmanuel Raj is a Finland-based Senior Machine Learning Engineer. He is a passionate ML Researcher, Software engineer, speaker, and author. He is also a Machine Learning Engineer at TietoEvry and a Researcher at Arcada University of Applied Sciences in Finland. With over 6+ years of experience building ML solutions in the industry, he has worked on multiple domains such as Healthcare, Manufacturing, Finance, Retail, e-commerce, aviation, etc.
Emmanuel is passionate about democratizing AI and bringing state-of-the-art research to the industry. He has a keen interest in R&D in technologies such as Edge AI, Blockchain, NLP, MLOps, and Robotics. He believes the best way to learn is to teach and he is passionate about teaching about new technologies, that's one reason for writing a book and making an online course on MLOps.
Emmanuel is the author of the book "Engineering MLOps". The book covers industry best case practices and hands-on implementation to Rapidly build, test, and manage production-ready machine learning life cycles at scale. There is a big evolution happening in Data science for good, and we are moving away from notebooks and models sharing to a collaborative way of working via MLOps. We will discuss this big evolution of DevOps, MLOps, Data Engineering, Data Science, and Data-Driven business in the meetup.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Emmanuel on LinkedIn: https://www.linkedin.com/in/emmanuelraj7/
// Other Links:
www.emmanuelraj.com
https://www.youtube.com/watch?v=m32k9jcY4pY
https://www.youtube.com/watch?v=1sGECHbc9zg
MLOps community meetup #68! Last Wednesday we talked to Veselina Staneva of TeachableHub, Simarpal Khaira of Intuit, and Korri Jones of Chick-fil-A, Inc.
//Abstract
Building, Designing, or even just casting the vision for MLOps for your company, whether a large corporation or an agile start-up up, shouldn't be a nigh-impossible task. Complex, but not an impossible mountain to climb.
In this meetup, we talked about the steps necessary to unlock the potential of data science for your organization, regardless of size.
//Bio
Veselina Staneva - Co-founder & Head of Product, TeachableHub
Over the past few years, Vesi work at a product company called CloudStrap.io, where together with her team they are simplifying cloud technologies and crafting modern solutions that lay a solid foundation for digital transformation at scale.
Vesi's main focus currently is their new product TeachableHub.com - an ML deployment and serving platform for teams, where she heads Product and Customer Development. In the past, Vesi had quite a diverse experience in managing projects for global enterprise companies such as telecommunications and internet service provider GTT and managed printing services giant HPInc, as well as deep-diving into e-commerce business development while running online stores on 7 Amazon markets as well as WordPress shops, where she managed to get from 0 to $30K MRR in less than a year without a dollar spent on paid advertising.
In Vesi's free time, she enjoys spending the rest of her energy doing all kinds of sports, as well as participate in non-professional triathlons and mountain bike ultra races.
Simarpal Khaira - Senior Product Manager, Intuit
Simarpal is the product manager driving product strategy for Feature Management and Machine Learning tools at Intuit. Prior to Intuit, he was at Ayasdi, a machine learning startup, leading product efforts for machine learning solutions in the financial services space. Before that, he worked at Adobe as a product manager for Audience Manager, a data management platform for digital marketing.
Korri Jones - Senior Lead Machine Learning Engineer, Chick-fil-A, Inc.
Korri Jones is a Sr Lead Machine Learning Engineer and Innovation Coach at Chick-fil-A, Inc. in Atlanta, Georgia where he is focused on MLOps. Prior to his work at Chick-fil-A, he worked as a Business Analyst and product trainer for NavMD, Inc., was an adjunct professor at Roane State Community College, and instructor for the Project GRAD summer program at Pellissippi State Community College and the University of Tennessee, Knoxville.
Korri's accolades are just as diverse, and he was in the inaugural 40 under 40 for the University of Tennessee in 2021, Volunteer of the year with the Urban League of Greater Atlanta with over 1000 hours in a single calendar year and has received the “Looking to the Future” award within his department at Chick-fil-A among many others, including best speaker awards in business case competitions. However, the best award he has received so far is being a loving husband to his wife Lydia.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vesi on LinkedIn: https://www.linkedin.com/in/veselina-d-staneva/
Connect with Simar on LinkedIn: https://www.linkedin.com/in/simarpal-khaira-6318959/
Connect with Korri on LinkedIn: https://www.linkedin.com/in/korri-jones-mba-780ba56/
Coffee Sessions #43 with Kyle Gallatin of Etsy, Maturing Machine Learning in Enterprise.
//Abstract
The definition of Data Science in production has evolved dramatically in recent years. Despite increasing investments in MLOps, many organizations still struggle to deliver ML quickly and effectively. They often fail to recognize an ML project as a massively cross-functional initiative and confuse deployment with production. Kyle will talk about both the functional and non-functional requirements of production ML, and the organizational challenges that can inhibit companies from delivering value with ML.
// Bio
Kyle Gallatin is currently a Software Engineer for Machine Learning Infrastructure at Etsy. He primarily focuses on operationalizing the training, deployment, and management of machine learning models at scale. Prior to Etsy, Kyle delivered ML microservices and lead the development of MLOps workflows at the pharmaceutical company Pfizer. In his spare time, Kyle mentors data scientists and writes ML blog posts for Towards Data Science.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Kyle on LinkedIn: https://www.linkedin.com/in/kylegallatin/
// Takeaways
Data science is still poorly defined and there is a large variance in organizational maturity
Basically, everything we need for mature ML in modern organizations exists technically except for the strategy, mentality, organization, and governance
Organizations who poorly define data science often overburden their data scientists, but there are expectations that data scientists know some engineering
Operationalizing data science is not that different from software engineering, and software engineering can be one of the most valuable skill sets for a data scientist.
// Q&A with Kyle as a data science mentor:
https://www.youtube.com/watch?v=7byRQGHD39w&t=1s
MLOps community meetup #66! Last Wednesday we talked to Alfredo Deza, Author and Speaker.
//Abstract
In this episode, the MLOps community talks about the importance of bringing DevOps principles and discipline into Machine Learning. Alfredo explains insights around creating the MLOps role, automation, constant feedback loops, and the number one objective - to ship Machine Learning models into production.
Additionally, we covered some aspects of getting started with Machine Learning that is critical, in particular how democratization ML knowledge is critical to a better environment, from libraries to courses, to production results. Spreading the knowledge is key!
//Bio
Alfredo Deza is a passionate software engineer, speaker, author, and former Olympic athlete. With almost two decades of DevOps and software engineering experience, he teaches Machine Learning Engineering and gives lectures around the world about software development, personal development, and professional sports.
Alfredo has written several books about DevOps and Python including Python For DevOps and Practical MLOps. He continues to share his knowledge about resilient infrastructure, testing, and robust development practices in courses, books, and presentations.
Alfredo Deza is the author of Python for DevOps and Practical MLOps.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Chris on LinkedIn: https://www.linkedin.com/in/chrisbergh/
Timestamps:
[00:00] Introduction to Alfredo Deza
[03:00] Alfredo's background in tech
[13:15] Who is this book for?
[14:15] "The reason why we need a Machine Learning book is that there's definitely a knowledge gap."
[16:05] Hierarchy of MLOps
[17:16] "Automation has to be the basis of pretty much, everything."
[19:03] Logging - "When in doubt, log it out!"
[24:50] Maturity
[29:55] "The notion of self-healing is very appealing."
[31:20] Learning Test
[37:40] "Catch things as early as possible. Anything that comes at the end of the process, the closer you are to the production, the more expensive it could get."
[37:54] "Expensive can be the dollar amount in engineering time, or it can be the dollar amount in services that you're using to produce, and the dollar amount on how long it would take to ship the version that fixes the problem."
[39:20] "Why not scan your containers before they hit the production and catch anything that has a critical vulnerability announced?"
[40:08] Interrupibility standards and pains
[42:34] "It is critical that we make it easier. How about we no longer point fingers and stigmatize people who don't do Machine Learning. The more people doing Machine Learning today, the better we're off."
[45:50] Simple and opinionated or flexible and complex
[46:45] "You have to strike a balance but you have to stay true to your principles."
[50:38] Abstraction Layers
[56:57] Take a risk or stay safe?
[57:20] "I think, you're gonna have risk everywhere you are. You're gonna have risk when you hire a Machine Learning Engineer. You're gonna have a risk with a Data Scientist. You're gonna have a risk with a Software Engineer."
MLOps community meetup #65! Last Wednesday we talked to Kseniia Melnikova, Product Owner (Data/AI), SoftwareOne.
//Abstract
In this MLOps Meetup, we talked about the Machine Learning model lifecycle and development stages and then analyze the main mistakes that everybody does at each stage. Kseniia also provided the audience with solutions to the mistakes and we discussed existing tools for experiment management.
//Bio
Kseniia is a product owner for Data/AI-based products. Right now, she is working mostly with numeric data analysis, customer insights, and product recommendations.
Previously Kseniia worked at Samsung Research with the biometrics team. She was studying computer science in Russia (Moscow) and a little bit of management in South Korea (Seoul). One of the most interesting directions of research - Model Lifecycle Management Systems and Reproducibility.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Kseniia on LinkedIn: https://www.linkedin.com/in/kseniia-melnikova/
Timestamps:
[00:00] Introduction to Kseniia Melnikova
[02:00] MLOps World Conference Announcement
[03:40] AI Development process: Common Mistakes
[07:45] Step 1: Planning
[07:48] Mistake #1: Personal Decisions - Teamwork
[08:31] Mistake #1: Cases
[09:00] Mistake #1: Solution
[11:52] Scrum
[12:50] "In Scrum, it's hard to plan because especially in research, you don't know which result affects new tasks that's why it might be a little slow for Machine Learning."
[14:28] Step 2: Data Processing
[14:34] Mistake #2: Chaos with Datasets
[15:26] Mistake #2: Cases
[16:48] Mistake #2: Solution
[20:12] Step 3: Experiments
[20:21] Mistake #3: Lack of Experiments Tracking
[22:13] Mistake #3: Case - Manual Experiments Tracking
[24:10] Mistake #3: Solutions
[25:57] Experiments Tracking Tools Example: MLFlow UI
[26:46] Awareness of Existing Tools
[28:21] Tools' Features
[29:21] Possible Combination
[29:48] Another Possible Combination
[30:24] Best Practice
[31:42] Mistake #0: Lack of Information Sharing
[32:26] Mistake #0: Solution - Organize more meetings/standups!
[34:18] Find Your Mistakes
[34:41] Mistake #0: Solution - Organize more meetings/standups!
[35:35] Audio Data
[39:32] Experiment tracking of only 1 ML engineer
[41:38] "I prefer reproducibility tools because it's automatic and it also takes a lot of time to manually upload the results into conference."
[43:03] AI Development Check-list
[43:40] Check-list Results
[44:52] "I think it's always interesting to rate yourself to share the results with other people to compete out of it."
[45:10] Why to Implement
[45:17] "If we have more automation on experimentations for data sets versioning, it will lead to less manual work."
[45:28] "AI Development process implementation will have the possibility to reproduce and compare experiments."
[45:37] "AI Development process implementation will make you comfortable on solving the issues you'll face every day."
[45:52] "AI Development process implementation will lead to a faster commercialization cycle because you will take less time on the process and more time for the results."
[46:03] "If we will take all the principles of AI Development process implementation, it will lead to easy communication between team members. You'll gain trust, have great teamwork, and everyone will have respect for each other."
[46:50] War stories prior to having AI Development process
[49:50] Calculating the lost money
Coffee Sessions #42 with Amit Paka of Fiddler AI, Model Performance Monitoring.
//Abstract
Machine Learning accelerates business growth but is prone to performance degradation due to its high reliance on data. Moreover, MLOps is often fragmented in many organizations, causing frictions to debug models in production. With new rules from the EU that focus on trust and transparency, it’s becoming more important to keep track of model performance. But how? We propose a new framework, a centralized ML Model Performance Management powered by Explainable AI. Learn more about how you can stay compliant while maximizing your model performance at all times with explainability and continuous monitoring.
//Bio
Amit is the co-founder and CPO of Fiddler, a Machine Learning Monitoring company that empowers companies to efficiently monitor and troubleshoot ML models with Explainable AI. Prior to founding Fiddler, Paka led the shopping apps product team at Samsung. Paka founded Parable, the Creative Photo Network, now part of the Samsung family. He also led PayPal's consumer in-store mobile payments launching innovations like hardware beacon payments and has developed successful startup products particularly in online advertising - paid search, a contextual, ad exchange, and display advertising. Paka has passions for actualizing new concepts, building great teams, and pushing the envelope, and aims to leverage these skills to help define how AI can be fair, ethical, and responsible.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Amit on LinkedIn: https://www.linkedin.com/in/amitpaka/
Timestamps:
[00:00] Thank you to Fiddler AI!
[00:46] Introduction to Amit Paka
[05:04] Amit's background in tech
[09:55] EU Regulation
[12:39] "The goal that the EU seems to be going for is they want to go for helping build human-centric and responsible AI."
[13:28] 4 AI Categories:
1. Unacceptable risk applications
2. High-risk applications
3. Limited risk applications
4. Minimal risk applications
[14:58] Deep dive into High-risk applications
[17:28] Digital Services Act (DSA) and Digital Marketing Act (DMA)
[19:02] Military
[19:33] "They don't know what they don't know and they probably wanted the door open."
[21:13] US on JIC Team - transparency and increasing trustworthiness on AI
[23:06] Diversity of industries and Explainability
[24:22] "The urgent need for Explainability comes from verticals that are facing the problems today on the ground and cannot run their business." [30:09] Model Performance Management (MPM)
[34:05] "When your model is facing issues, you now have to root-cause it within life."
[35:40] Control Theory
[36:10] "Control Theory means that you do not just measure it but you can influence it so you can actually keep it."
[38:14] Abstraction into being useful
[43:23] "You can train a model that accurately represents the reality."
[44:00] Data scientist doing ML Flow
[49:55] Amit's favorite surprise!
[53:04] Banking and Insurance adoption of ML
[55:48] Advise ML Scientists and Data Scientists in terms of Explainable AI
[58:25] "Models are incredibly hard to debug. You're just training a model for high accuracy but you don't know how that accuracy is distributed."
[59:49] Linking of EU Regulation and MPM
Coffee Sessions #41 with Monmayuri Ray of Gitlab, CI/CD in MLOPS.
//Abstract
We all are familiar with the concept of MVP. In the world of DevOps, one is also familiar with Minimal Viable Feature and further Minimal Viable change. CI/CD is the orchestrator and the underlying base to enable automated experimentation, to start small, and build an idea for production. Now if we use the same fundamentals in MLOps, what does that mean?
The podcast will take the audience on a journey in understanding the fundamentals of orchestrating machine predictions using responsible CI/CD in MLOps in this ever-changing, agile world of software development. One shall hope to learn how to excel at the craft of CI for Machine Learning (ML), lowering the cost of deployment through a robust CI/CD/CT/CF framework.
//Bio
Monmayuri is an advisor, data scientist, and researcher specializing in MLops/DevOps at GitLab in Sydney. She builds creative, products to solve challenges for companies in industries as diverse as financial services, healthcare, and human capital.
Along the way, Mon has built expertise in Natural Language Processing, scalable feature engineering, MLOps transformation and digitization, and the humanization of technology. With a background in applied mathematics in biomedical engineering, she likes to describe the essence of AI as “low-cost prediction” and MLOps as “low-cost transaction” and believes the world needs the collaboration of poets, historians, artists, psychoanalysts and scientists, engineers to unlock the potential of these emerging technologies where one works in making a machine think like humans and be efficient automated fortune tellers.
//Takeaways
Key Takeaways include how to incorporate the best CI/CD practice in your MLOPS lifecycle. Things to do and things not to do. How best to get the DevOps engineer, ML engineer, and data scientists to speak the same language and automate CI for pipeline and models.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Mon on LinkedIn: https://www.linkedin.com/in/monmayuri-ray-713164a0/
Timestamps:
[00:00] Introduction to Monmayuri Ray
[00:57] Mon's background in tech
[02:50] MLOps being approached at Gitlab
[07:00] CI/CD for MLOPS Definition
[07:57] "AI is the dropping cost of machine prediction."
[10:25] MLOps and other tools fitting into Gitlab
[12:18] "If you want to have an MLOps first strategy, anything you are putting first needs to be substituted with what you had before first. It's really important then to know your priorities."
[15:24] Process of how to build
[18:16] "Before getting into even understanding the maturity, understand the outcome."
[18:45] Challenges in CI/CD for MLOps
[19:50]" Automation also empowers collaboration."
[24:15] Keeping up
[28:33] "I think, the best tools and frameworks are to give people the freedom to be the best version of who they are. As a system, being governed, having that controlled freedom, you can be more Human."
[31:20] Resources to suggest in terms of MLOps Education
[32:12] Understand the business outcomes of MLOps - Understanding the economics of AI and Machine Learning - Cultural shift
[35:57] Effectiveness of understanding the business outcomes of MLOps to Gitlab customers.
[39:42] "It's judgment, action, outcome, and how does this fully impact the overall workflow."
[40:00] Enabling vs Keeping the guardrails on
[43:26] Best practices
MLOps community meetup #64! Last Wednesday we talked to Christopher Bergh, CEO, DataKitchen.
//Abstract
Working on a shared technically difficult problem there will be some things that are important no matter what industry you are in. Whether it's building cars in a factory, using agile or scrum methodology, or productionizing ML models you need a few basics. Chris gives us some of his best practices in the conversation.
//Bio
Chris Bergh is the CEO and Head Chef at DataKitchen. Chris has more than 25 years of research, software engineering, data analytics, and executive management experience. At various points in his career, he has been a COO, CTO, VP, and Director of Engineering. Chris is a recognized expert on DataOps. He is the co-author of the "DataOps Cookbook” and the “DataOps Manifesto,” and a speaker on DataOps at many industry conferences.
//Takeaways
Your model is not an island. For success, Data science requires a high level of technical collaboration with other parts of the data organization.
//Other Links
On-Demand Webinar - Your Model is Not an Island: Operationalizing Machine Learning at Scale with ModelOps
https://info.datakitchen.io/watch-on-demand-webinar-operationalize-machine-learning-at-scale-with-modelops
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Chris on LinkedIn: https://www.linkedin.com/in/chrisbergh/
Timestamps:
[00:00] Introduction to Christopher Bergh
[02:57] MLOps community in partnership with MLOps World Conference
[04:34] Chris' Background
[07:59] "When we started with the company, I realized that the problem I have is generalizable to everyone. I'm getting enough there in years and I wanted to remove the amount of pain that other people have."
[09:53] DataOps vs MLOps
[10:15] "I don't really honestly care what Ops you use, right? Hahaha! Call it your favorite Ops 'cause first of all as an engineer, I want precise definitions. I look at it from a completely odd-ball way so you could call it whatever Ops term you want."
[12:45] Best practices of companies
[14:16] "When that code runs in production, monitor and check to see if it's right. Absorb it, monitor it because the model could go out of tune. The data going into it could be wrong. The data transformation could break. Shit happens and don't trust your data providers."
[19:00] The whole is still greater than its part
[20:26] "It is harder to focus on the results than just under a piece of the task. Don't spend too much time on doing the wrong thing."
[23:50] DevOps Principles and Agile
[27:17] DataOps Manifesto - DataOps is Data Management reborn
[27:45] "The 'Ops' term is ending up encompassing the work that you do in addition to the system you build to do the work."
[30:45] Standardization
[32:22] "I think that there's a lack of perception of the need to spend time on doing the operations part of the equation."
[34:15] Tools as lego blocks
[34:49] "Good interphases make good neighbors."
[36:23] "Standards can help but they're not the panacea."
[36:30] Cultural side - You build it, you own it, you ship it
[39:28] Value chain
[44:19] Ripple effect of testing
[48:03] Google on "One tool to rule them all"
[49:50] "Legacy happens if you're gonna live in the real world and not start greenfield projects."
[53:47] Starting MLOps in the legacy system
Coffee Sessions #40 with Srivatsan Srinivasan of AIEngineering, Scaling AI in Production.
//Abstract
//Bio
20+ years of intense passion for building data-driven applications and products for top financial customers. Srivatsan has been a trusted advisor to a senior-level executive from business and technology, helping them with complex transformation in the data and analytics space. Srivatsan also run a YouTube Channel (AIEngineering) where he talks about data, AI and MLOps.
//Takeaways
Understand the role and need of MLOps
Prioritize MLOps capability
Model deployment
Importance of K8s
//Other Links
AI and MLOps free courses - https://github.com/srivatsan88
Youtube channel: bit.ly/AIEngineering
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Srivatsan on LinkedIn: https://www.linkedin.com/in/srivatsan-srinivasan-b8131b/
Timestamps:
[00:00] Introduction to Srivatsan Srinivasan
[01:41] Background on Youtube AIEngineering
[03:17] Tips on learning MLOps and start with the field
[06:00] "Focus on your key challenges and that will drive your capability that you need to implement."
[06:50] Tips on starting CI/CD
[08:46] "Start with DevOps and see what additional capabilities you will require for the Machine Learning aspect of it."
[09:24] Staying general in different environments
[10:43] "Focus on the core concepts of it. The concepts are similar."
[12:10] Testing systems robustly
[20:00] Trends within MLOps space
[20:31] "Everybody can fail fast but you need to fail smart because Machine Learning is a huge investment."
[23:21] GCP Auto ML
[26:54] Deployment
[27:06] "It's not only the tools, but it's also the patterns."
[29:34] Kubernetes perspective
[31:21] Favorite model release strategy
[36:22] Annotation, labeling, and concept of ground truth
[38:10] Best practices in Architecture and systems design in the context of ML
[41:29] "You learn a lot, at the same time the complexity also increases, so work with multiple teams in this process to learn it."
[42:35] "Your speed increases based on the way you envision your architecture."
[42:55] Software engineering lifecycle vs machine learning development life cycle
[44:55] Youtube experience
[45:50] "My focus has always been from intermediate to experts."
[46:24] Content creation
[47:17] "You cannot do everything in MLOps at one stretch. You have to see what is critical for you."
[47:23] "For me, continuous training is not that critical because I don't want to take the freedom out of the data scientists."
[48:31] New contents planned
[48:40] IoT and Edge Analytics - Predictive maintenance
[50:21] "It's a two-way process. I learn then I teach."
Coffee Sessions #39 with Stephen Galsworthy of Quby, MLOps: A leader's perspective.
//Abstract
//Bio
Dr. Stephen Galsworthy is a data leader skilled at building high-performing teams and passionate about developing data-powered products with lasting impact on users, businesses, and society.
Most recently he was the Chief Data and Product Officer at Quby, an Amsterdam-based tech company offering data-driven energy services. He oversaw its transformation from a hardware-based business to a digital organization with data and AI at its core. He put in place a central cloud-based data infrastructure and unified analytics platform to collect and take advantage of petabytes of IoT data. His team deployed real-time monitoring and energy insight services for 500k homes across Europe.
Stephen has a Master’s degree and Ph.D. in Mathematics from Oxford University and has been leading data science teams since 2011.
//Takeaways
MLOps as a process, people, and technological problem.
Experiences from a team working at the forefront of data and AI.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Stephen on LinkedIn: https://www.linkedin.com/in/galsworthy/
MLOps community meetup #63! Last Wednesday we talked to Felipe Campos Penha, Senior Data Scientist, Cargill.
//Abstract
Can one learn anything useful by creating content online? The usual answer is a sounding YES. But what about live coding an MLOps project on Twitch? Can anything good come out of it?
//Bio
Felipe Penha creates content about Data Science regularly on the Data Science Bits channel on YouTube and Twitch. He has 8+ years of experience with hands-on data-related work, starting with his doctorate in Astroparticle Physics. His career in the private sector has been devoted to bringing value to various segments of the Food and Beverages Industry through the use of Analytics and Machine Learning.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Felipe on LinkedIn: https://www.linkedin.com/in/fcpenha/
Timestamps:
[00:00] Introduction to Felipe Campos Penha
[01:30] Felipe's background
[05:36] Developing models in physics vs developing models for companies
[08:07] Felipe's transition from Jupyter Notebook to Operational ML
[09:34] "The thought of business basically for customers, they always wanted to see the value and try to roll out more manual work like spreadsheets so they could try out that model on the field."
[12:07] Felipe's software engineering development learning
[14:10] Catalyst on Youtube and Twitch
[18:06] Elements of Twitch
[20:02] Non-polished versions of Twitch
[21:16] "Twitch was not made for coding, it was for gamers."
[26:17] Felipe's audience impact on Twitch
[28:02] Logistical pieces
[30:43] Words of wisdom on live streaming
[30:56] "Don't be afraid to start. There are many streamers that are actually learning from scratch and they are showing the process of learning online. They are learning faster because the help is faster."
[33:16] Blog post as other means to Twitch
[33:50] "I'm a perfectionist when I'm writing. The shortest it is, the hardest it could get. You want to polish it a lot to make nice figures. I learned a lot but for me, I feel that process is too slow because you're thinking about one subject for a long time trying to polish it while in live streaming, it's very dynamic and fast."
[34:25] Twitch affecting Felipe's career
[36:36] "Exposing yourself, showing your mistakes, vocalizing your thoughts, I think all of this makes you a better programmer."
[37:12] Getting through a problem
[39:41] Recommended streamers that caught Felipe's interest
[41:00] Community aspect and importance of Twitch
[42:42] Role of community on Twitch
[45:16] "Twitch is becoming such a trend that even companies are following."
MLOps community meetup #62! Last Wednesday we talked to Oguzhan Gencoglu, Co-founder & Head of AI, Top Data Science.
//Abstract
Starting the AI adoption with AI Proof-of-Concepts (PoCs) is the most common choice for most companies. Yet, a significant percentage of AI PoCs do not make it into production whether they were successful or not. Furthermore, running yet another AI PoC follows the law of diminishing returns in various aspects. This talk will revolve around this theme.
//Bio
Oguzhan "Ouz" Gencoglu is the Co-founder and Head of AI at Top Data Science, a Helsinki-based AI consultancy. With his team, he delivered more than 70 machine learning solutions in numerous industries for the past 5 years. Before that, he used to conduct machine learning research in several countries including the USA, Czech Republic, Turkey, Denmark, and Finland. Oguzhan has given more than 40 talks on machine learning to audiences of various backgrounds.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Oguzhan on LinkedIn: https://www.linkedin.com/in/ogencoglu/
Timestamps:
[00:00] Introduction to Oguzhan Gencoglu
[00:47] Ouz's background
[01:47] Recurring/repetitive problem patterns
[03:16] "When you solve a repetitive task in an automatic way, that's Scalability."
[04:32] Evolution expected of Machine Learning
[05:10] "People are quite confused about the titles and what's worst, those titles don't have a common definition in different companies. If you feel a little bit overwhelmed, that's normal."
[08:04] Proof-of-Concepts
[10:35] Successful PoCs but not Productionized
[16:03] Productionize as soon as possible
[16:47] "In your Proof-of-Concepts, it's not only technical, but it's also a mindset."
[20:00] Framework of a successful PoCs
[24:28] Taking too much on PoCs
[28:05] Proof-of-Concepts after Proof-of-Concepts and Proof-of-Concepts hell
[31:30] Wholistic view
[34:00] Operationalizing PoCs
[37:17] "The teams also need to adjust themselves to these new tools, new paradigms, and the different needs of the whole industry."
[37:26] Horror stories
[39:54] Open communication tips
43:31] "Open communication should not only be from the technical perspective but also down to the business and strategy perspective."
[44:20] Translation tips
[44:39] "I believe the most crucial part of today's ML scientists' role is not building a machine learning model but translating a real-life problem into a machine learning problem. It's crucial because it's a scarce talent and skill."
[49:30] Realistic budget for small PoCs
[50:18] "You need at least 1 month of work of proof of value but that doesn't mean things will go to production."
[51:40] Understanding the questions fully
[52:55] "That translation skill is the greatest skill to have in this industry because you can't auto ML that or whatever. It stands the test of time because that will be needed all the time."
Coffee Sessions #38 with Adam Sroka of Origami Energy, Organisational Challenges of MLOps.
//Abstract
Deploying data science solutions into production is challenging for both small and large organizations. From platform and tooling wars to architecture and design pattern trade-offs it can get overwhelming for inexperienced teams. Furthermore, many organizations will only go through the painful discovery process once. Adam will share some of his experiences from consulting and leading data teams to successfully deploying machine learning solutions, highlighting some of the more difficult challenges to overcome. You might not be surprised to hear it’s not all down to the tech.
//Bio
Dr. Adam Sroka, Head of Machine Learning Engineering at Origami Energy, is an experienced data and AI leader helping organizations unlock value from data by delivering enterprise-scale solutions and building high-performing data and analytics teams from the ground up. Adam shares his thoughts and ideas through public speaking, tech community events, on his blog, and in his podcast.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Adam on LinkedIn: https://www.linkedin.com/in/aesroka/
Timestamps:
[00:00] Introduction to Adam Sroka
[01:53] Adam's background in tech
[08:06] 2 blog posts of Adam: Why So Many Data Scientists Quit Good Jobs at Great Companies and Why You Shouldn’t Hire More Data Scientists [08:31] High turn rate Adam has in the data science role
[13:50] Avoiding hiring talents with deficits and coaching people
[16:05] "I can't teach you to care about the standard of your core of what you're doing. It's quite hard to teach people charisma. Everything else, you pick up."
[16:45] Resume-driven development, the idea of not playing the game, and politics in the workplace.
[17:57] "You have to realize, other people, don't have the same experience in the context that you do."
[19:59] Exit, Voice, Loyalty and Neglect Model
[22:35] You probably don't need a data scientist
[23:40] "Data scientists can do everything slower and more expensively than everyone else but they can do everything and that's the important bit."
[27:54] "My success is just driven by who I am as much as what I can do." Vishnu
[28:24] Being Candor
[30:37] Disconnect between the senior stakeholders and data scientists
[32:30] "Before you come out to bring someone in some expensive talent search, engage with the consultancy. Do a four-week PRC, get them to tell you like."
[34:18] Educational experiences as a consultant
[37:35] Adam's journey into MLOps, productionize ML models when you are a data scientist and tips
[43:16] "Beginners can help beginners. Your perspective is really important. The value is not in the content. The value is in your perspective of the content."
[45:21] Educating clients on uncertainty
[48:34] Decision making process
[52:32] Organizational problems
[53:43] "All models are wrong, but some are useful." George Box
MLOps community meetup #61! Last Wednesday we talked to Lex Beattie, Michael Munn, and Mike Moran.
//Abstract
We started out talking about some of the main bottlenecks they have encountered over the years of trying to push data products into production environments. Then things started to heat up as we dove into the topic of monitoring ML and inevitably the word explainability started being thrown around.
Turns out Lex is currently doing a Ph.D. on the subject so there was much to talk about. We had to ask if explainability is now table-stakes when it comes to monitoring solutions on the market? The short answer from the team. Yes!
Please excuse the bit of sound trouble we had with google mike at the beginning.
//Bio
Lex Beattie - ML Engagement Lead, Spotify
In the last year, Lex has helped over 40 different teams across Spotify understand ML best practices, productionize ML workflows and implement impactful ML in their products. Lex is also a Ph.D. candidate at the University of Oklahoma, focusing on feature importance and interpretability in deep neural networks. Beyond her passion for all things ML, she enjoys exploring the great outdoors in Montana with her German Wirehaired Pointer, Bridger.
Michael Munn - ML Solutions Engineer, Google
Michael is an ML Solutions Engineer with Google Cloud and Google's Advanced Solutions Lab. In his role, he works with customers to build and deploy end-to-end ML solutions with Google Cloud. Within the Advanced Solutions Lab, he teaches these skills to customers.
Mike Moran - Principal Engineer, Skyscanner
Mike has worked across many dimensions; in large/tiny companies, back-end/front-end, with many languages, and as a sys-admin /engineer/manager. Mike has a healthy skepticism for most things and likes solving problems through applying System thinking.
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Lex on LinkedIn: https://www.linkedin.com/in/lexbeattie/
Connect with Michael on LinkedIn: https://www.linkedin.com/in/munnm/
Connect with Mike on LinkedIn: https://www.linkedin.com/in/mrmikemoran/
Timestamps:
[00:00] Introduction to Lex, Michael, & Mike
[02:46] Common roadblocks
[05:25] Consolidating knowledge
[07:02] Bottlenecks on failures
[09:58] Don't go on a detour
[12:22] Bringin on complexity signs
[19:33] Explainable AI
[21:34] "There are different ways to approach Explainable AI. It starts to get more complicated when you start working with more complicated models." Lex
[24:43] "If there are a lot of disparate sources out there about Explainability, I'd found myself hunting down various resources to simplify it for customers I'd worked with." Michael
[26:46] "Being clear about who you're explaining it to because in our context, sometimes the organization needs to explain it to a regulator." Mike [28:04] Monitoring solution
[31:00] ML Canvas
[33:24] Explainable AI Resources
[34:48] Explainable Predictions by Michael
[36:48] Purpose of Explainable Model
[39:40] Work in the same language
[42:46] Use of War Stories
[49:11] Hot seat!
[49:15] Mike - Skyscanner pricing
[50:30] Lex - Spotify recommendation sudden stop
[51:35] Michael - NLP models on emails
Coffee Sessions #37 with Ariel Biller of ClearML, MLOps Memes.
//Abstract
The Meme king of MLOps joins us to talk about why we need more MLOps memes and how he got so damn good at being able to zoom out and see things from a metta level them make a meme about it!
//Bio
A researcher first, developer second, in the last 5 years Ariel worked on various projects from the realms of quantum chemistry, massively parallel supercomputing, and deep-learning computer vision. With AllegroAi, he helped build an open-source R&D platform (Allegro Trains), and later went on to lead a data-first transition for a revolutionary nanochemistry startup (StoreDot). Answering his calling to spread the word on state-of-the-art research best practices, He recently took up the mantle of Evangelist at ClearML. Ariel received his Ph.D. in Chemistry in 2014 from the Weizmann Institute of Science. With a broad experience in computational research, he made the transition to the bustling startup scene of Tel-Aviv, and to cutting-edge Deep Learning research.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Ariel on LinkedIn: https://www.linkedin.com/in/LSTMeow/
//Other Links
https://youtu.be/1C_l5ICJlEo
https://youtu.be/yTtTrwXEhN4
https://youtu.be/F4Ghp-phFuI
Timestamps:
[00:00] Introduction to Ariel Biller
[01:20] Ariel's background
[03:40] Story behind Memeing
[06:36] "Memes can be as extreme as you want because people don't know if they're going to take you seriously or you're joking."
[07:21] MLOps memes and more
[10:15] MLOps fear
[13:00] MLOps being more complicated than DevOps.
[13:10] "A meme material is a social commentary about what there is and what there is now."
[16:00] Standardization
[18:18] "Would we have MLOps' code in a sweeping way or not?"
[18:26] "I'm not sure as a community of builders, we have the right perspective that will walk for all the cases."
[20:26] Journey into evangelism
[26:45] "Feature stores are a big meme."
[27:08] "Memeing is like a muscle. If you flex it daily it creates tensions."
[31:26] We need to de-jargonize MLOps and ML engineering
[35:55] Current Israeli tech scene
[39:16] "The deficit is that there's a limited amount of people doing MLOps right now."
[43:14] Tooling space
[46:57] "Concentrate on the basic stuff that will survive forever and if you need to reach out for a tool, don't reach out for a tool, reach out for obstruction."
[51:47] Standardization of ID Tree
[52:43] "Everybody is doing whatever they want because it works for them. Someday, someone would come out with some good obstruction and good toolchain that works across the board that will click for everyone and will use it from that time on."
[55:20] Ecosystem support
Coffee Sessions #36 with Luigi Patruno of 2U, Luigi in Production Part 2.
//Abstract
Learning Voraciously: We talk a lot in the community about how to learn and upskill in an efficient way. Luigi provided great insight into how he applies certain principles to his learning practices. One tip he shared is to rigorously read and digest books. Luigi himself has used books to address his knowledge gaps in areas like product, finance, etc. I appreciated the emphasis on books. A lot of the reason we feel inundated by new learning resources is that they are online. Emphasizing books, which are often far higher-quality than blog posts, can slow things down and focus our learning.
Leadership Patience: Lately, Luigi has been spending more time managing projects and the data science team at 2U. He shared a lot of his insights into how to manage data science and machine learning properly. One of the most important things he emphasized to us was his patient attitude towards solving problems important to leadership. Turning around organizations is hard work. It's slow, it takes energy, and it is a nonlinear process. As he has course-corrected at various times as a data science leader, Luigi has brought admirable patience to the task, which has helped him be more successful on the things that matter to the entire company.
Communication Flows: It's easy to imagine Luigi as a great communicator, given his experience running MLInProduction.com. In our conversation, he showed us how he puts it to use in his management style. Luigi shared the importance of understanding how communication flows across an organization. Being aware of this is crucial to working on the right, most impactful things. Having a pulse on what different groups and leaders are thinking about can help you evaluate your impact as a team.
//Bio
Luigi Patruno is a Data Scientist focused on helping companies utilize machine learning to create competitive advantages for their business. As the Director of Data Science at 2U, Luigi leads the development of machine learning models and MLOps infrastructure for predicting student success outcomes across 2U’s portfolio of university partners. As the Founder of MLinProduction.com, Luigi creates and curates content to educate machine learning practitioners about best practices for running resilient machine learning systems in production. Luigi has consulted on data science and machine learning at Fortune 500 companies and start-ups and has taught graduate-level courses in Statistics and Big Data Engineering. He has an M.S. in Computer Science and a B.S. in Mathematics.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Luigi on LinkedIn: https://www.linkedin.com/in/luigipatruno91/
Timestamps:
[00:00] Introduction to Luigi Patruno
[01:12] Update about Luigi
[04:08] Luigi's transition
[07:18] Problem-focused
[11:00] New problem
[12:51] Rational platform strategy
[18:18] Bringing the learnings to the team
[20:57] Formulating and communicating vision
[25:40] Problem-driven mindset
[35:53] Organizational blind spots
[41:12] Continous learning
[42:46] "Default to reading."
[44:44] The Lindy effect
[46:20] "You'll fail less often on the easy problems."
[46:25] Act upon reading
[51:48] Ethical implications of ML
[53:24] "Machine Learning is predicated on leveraging data to uncover insights that went to otherwise be able to be uncovered."
Coffee Sessions #35 with Nick Masca of Marks and Spencer, War Stories Productionising ML.
//Abstract
A conversation with MLOps war stories. Better said, a war story conversation. The kind that informs modern MLOps best practices.
Nick shared how to make MLOps organizational changes at large companies. I loved one tidbit he mentioned--"it's an evolution, not a revolution". That's a frank observation about the speed of practical change. As we all know it doesn't happen overnight.
Another great learning Nick shared focused on the value of delivering incremental results regularly. Oftentimes, ML projects suffer because of a focus on delivering too much too soon. This can then lead to a trough of disappointment with the way things actually pan out. Nick shared his experience on how to avoid such pitfalls with us so you don't have to learn the hard way.
//Bio
Nick currently serves as a Head of Data Science at Marks and Spencer, a large retailer based in the UK. With a background originally in statistics, he transitioned into data science in 2014 and has picked up many battle scars and learnings since.
//Link to the MLOps War Stories
https://www.linkedin.com/posts/dpbrinkm_what-is-your-mlops-war-story-activity-6772604800971370496-LxtX
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Nick on LinkedIn: www.linkedin.com/in/nick-masca-09454956/
Timestamps:
[00:00] Introduction to Nick Masca
[01:36] Nick's background in tech
[05:01] Nick's current job
[06:19] Building the basics
[08:18] "If you can gain trust and demonstrate value early, you could also freeze you up to the tidy marks later."
[09:19] Strategy on long-running vision
[10:25] "Historically, the legacy waterfall processes in the business where teams have specialist responsibilities."
[11:14] KPI's
[12:36] KPI translations into action plans
[15:43] Data scientists call
[17:13] Nick's nightmarish story
[22:52] Making the case on such a nightmarish story
[25:06] Tools used by Marks and Spencer in 2015
[27:15] More complicated process
[28:08] Takeaways from experience
[30:57] Obstacles in deploying
[34:53] Simplifying models
[37:31] Combining environments into one
[38:45] "Having written standards can be quite helpful to take ownership and responsibility around that."
[40:23] M&S team interaction
[41:31] "It's an evolution, it's not a revolution I'd say at the moment but there's definitely real emphasis where we are to improve things and work towards goals to enable our team to work quicker, empower them."
[42:10] Team moralizing
[43:11] Takeaways from war stories
[43:30] "The biggest takeaway for me is to start small, keep things simple, try things and it can be surprising sometimes what you'll find. Something simple can give you surprising results."
[44:35] Opinions on Data Science and Machine Learning businesses democratize and commoditize
MLOps community meetup #60! Last Wednesday we talked to Vishnu Prathish, Director Of Engineering, AI Products, Innovyze.
//Abstract
The way Data Science is done is changing. Notebook sharing and collaboration were messy and there was minimal visibility or QA into the model deployment process. Vishnu will talk about building an ops platform that deploys hundreds of models at-scale every month. A platform that supports typical features of MLOps (CI/CD, Separated QA, Dev and PROD environment, experiments tracking, Isolated retraining, model monitoring in real-time, Automatic Retraining with live data) and ensures quality and observability without compromising the collaborative nature of data science.
//Bio
With 10 years in building production-grade data-first software at BBM & HP Labs, I started building Emagin's AI platform about three years ago with the goal of optimizing operations for the water industry. At Innovyze post-acquisition, we are part of the org building world-leading water infrastructure data analytics product.
//Takeaways
Why is MLOps necessary for model building at scale?
What are various cloud-based models for MLOps?
Where can ops help in various points in the ML pipeline Data Prep, Feature Engineering, Model building, Training, Retraining, Evaluation and inference
----------- Connect With Us ✌️------------- Join our Slack community:
https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vishnuprathish/
Timestamps:
[00:00] Introduction to Vishnu Prathish
[00:16] Vishnu's background
[04:18] Use cases on wooden pipes for freshwater
[04:55] Virtual representation of actual, physical, tangible assets
[06:56] Platform built by Vishnu
[08:30] Build a reliable representation of network
[11:52] Pipeline architecture
[16:17] "MLOps is still an evolving discipline. You need to try and fail many times before you figure out what's right for you."
[17:11] Open-sourcing
[18:17] Platform for virtual twin
[20:02] Entirely Amazon Stagemaker
[20:43] Data quality issues
[23:21] Reproducibility
[23:40] "Reproducibility is important for everybody. Most of the frameworks do that for you."
[25:00] Reproducibility as Innovyze's core business.
[26:38] Each model is individual to each customer
[27:50] Solving reproducibility problems
[28:24] "Reproducibility applies to the process of training pipelines. It starts with collected from historical raw data from customers. In real-time, there's also this data being collected directly from sensors coming from a certain pipeline."
[31:55] "Reusable training is step one to attaining automated retraining."
[32:17] Collaboration of Vishnu's team
[36:23] War stories
[41:36] Data prediction
[44:24] "A data scientist is the most expensive hire you can make."
[47:55] 3 Tiers
[48:53] MLOps problems
[52:25] Automatically retraining
[52:34] "Because of the numbers of models that go through this pipeline, it's impossible for somebody to manually monitor and retrain as necessary. It's not easy, it takes a lot of time."
[54:22] Metrics on retraining
[56:42] "Retraining is a little less prevalent for our industry compared to a turned prediction model that changes a lot. There are external factors that depend on it but a pump is a pump."
Coffee Sessions #34 with Geoff Sims of Atlassian, Machine Learning at Atlassian.
//Abstract
As one of the world's most visible software companies, Atlassian's vast data and deep product suite pose an interesting MLOps challenge, and we're grateful to Geoff for taking us behind the curtain.
//Bio
Geoff is a Principal Data Scientist at Atlassian, the software company behind Jira, Confluence & Trello. He works with the product teams and focuses on delivering smarter in-product experiences and recommendations to our millions of active users by using machine learning at scale. Prior to this, he was in the Customer Support & Success division, leveraging a range of NLP techniques to automate and scale the support function.
Prior to Atlassian, Geoff has applied data science methodologies across the retail, banking, media, and renewable energy industries. He began his foray into data science as a research astrophysicist, where he studied astronomy from the coldest & driest location on Earth: Antarctica.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Geoff on https://www.linkedin.com/in/geoff-sims-0a37999b/
Timestamps:
[00:00] Introduction to Geoff Sims
[01:20] Geoff's background
[04:00] Evolution of ML Ecosystem in Atlassian
[06:50] Figure out by necessity
[08:47] Machine Learning not priority number one and disconnected to MLOps
[11:53] Atlassian being behind or advanced?
[16:38] Serious switch of Atlassian around machine learning
[17:47] What data org did it come from?
[20:00] Consolidation of the stack
[21:21] Tooling - blessing and curse
[24:37] Tackling play out
[29:38] Staying on the same page
[30:48] Priority of needs
[31:55] How did it evolve?
[35:12] Where is Atlassian now?
[40:21] "Architecturally, Tecton is very very similar (to ours), it was just way more mature."
[41:17] What unleashed you to do now?
[41:36] "The biggest thing is independence from a data science perspective. Less reliance and less dependence on an army of engineers to help deploy features and models."
[44:25] Have you bought other tools?
[45:43] "At any given time, there's something that's a bottleneck. Look where the bottleneck is, then fix it and move on to the next thing."
[48:20] Atlassian bringing a model into production
[50:01] "When we undertake whatever the project is, its days or weeks to go to a prototype rather than months or quarters."
[53:10] "Conceptually, you're struggling walking towards that place because that's the place you want to be. If that's your problem, that's good. That's the promised land."
[54:45] "Using our own tools is paramount because we are customers as well. So we see and feel the pain which helps us identify the problems and understand them."
MLOps community meetup #59! Last Wednesday was the celebration of the MLOps Community 1 Year Anniversary! This has been a conversion of Demetrios Brinkmann, David Aponte and Vishnu Rachkonda!
//Abstract
Over the past year Demetrios, David and Vishnu have interviewed many of the top names in MLOps. During this time they have been able to apply these learnings at their jobs and see what works for them. In this one year anniversary meetup the three of them will discuss some of the most impacting advice they have received in the last year and how they have put it into practice.
//Bio
Demetrios Brinkmann
At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter.
David Aponte
David is one of the organizers of the MLOps Community. He is an engineer, teacher, and lifelong student. He loves to build solutions to tough problems and share his learnings with others. He works out of NYC and loves to hike and box for fun. He enjoys meeting new people so feel free to reach out to him!
Vishnu Rachakonda
Vishnu is the operations lead for the MLOps Community and co-hosts the MLOps Coffee Sessions podcast. He is a machine learning engineer at Tesseract Health, a 4Catalyzer company focused on retinal imaging. In this role, he builds machine learning models for clinical workflow augmentation and diagnostics in on-device and cloud use cases. Since studying bioengineering at Penn, Vishnu has been actively working in the fields of computational biomedicine and MLOps. In his spare time, Vishnu enjoys suspending all logic to watch Indian action movies, playing chess, and writing.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Timestamps:
[01:07] Big shoutout to everybody that's in these meetups!
[02:03] Big shoutout to Ivan Nardini for leading the Engineering Labs and to everyone who took part in the Engineering Labs!
[02:26] Big shoutout to Charlie You leading the Reading Group and to everyone who takes part in it!
[02:39] Big shoutout to everyone who takes part in the Office Hours!
[02:49] Big shoutout to the people who are helping with shaping the website!
[03:34] Thanks to all the people in Slack! Laszlo, Ariel, and people answering Slack questions.
[04:10] Big thanks to all our Sponsors FiddlerAI, Algorithmia, and Tecton!
[06:13] David's Background
[08:08] Vishnu's Background
[09:55] High-Level Points
[15:57] Starting small
[24:05] Over-optimization - the root of all evil
[26:42] Keeping text deck open
[36:45] Missing from current MLOps tooling
[48:00] How to communicate in these data products?
Coffee Sessions #33 with Sarah Catanzaro of Amplify Partners, MLOps Investments.
//Bio
Sarah Catanzaro is a Partner at Amplify Partners, where she focuses on investing in and advising high potential startups in machine intelligence, data management, and distributed systems. Her investments at Amplify include startups like RunwayML, Maze Design, OctoML, and Metaphor Data among others. Sarah also has several years of experience defining data strategy and leading data science teams at startups and in the defense/intelligence sector including through roles at Mattermark, Palantir, Cyveillance, and the Center for Advanced Defense Studies.
//We had a wide-ranging discussion with Sarah, three takeaways stood out:
// Other Links
https://amplifypartners.com/team/sarah/
https://projectstoknow.amplifypartners.com/ml-and-data
https://twitter.com/sarahcat21/status/1360105479620284419
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Sarah on LinkedIn: https://www.linkedin.com/in/sarah-catanzaro-9770b98/
Timestamps:
[00:00] Introduction to Sarah Catanzaro
[02:07] Sarah's background in tech
[06:00] Staying engineer oriented despite being an investment firm
[08:50] Tools you wished you had earlier in your career
[12:36] 2 Motives of ML Engineers and ML Platform Team
[16:36] Open-sourcing
[21:29] Startup focus on resources
[23:57] Playout of open-source project
[27:32] Consolidation
[33:18] Finding solutions
[36:18] Evolution of MLOps industry in the coming years
[42:36] Frameworks
[43:14] Structure data sets available to researchers. Meaningful advances of deep learning applied to structure data as well.
MLOps community meetup #58! Last Wednesday we talked to Ben Wilson, Practice Lead Resident Solutions Architect, Databricks.
Model Monitoring Deep Dive with the author of Machine Learning Engineering in Action. It was a pleasure getting to talk to Ben about difficulties in monitoring in machine learning. His expertise obviously comes from experience and as he said a few times in the meetup, I learned the hard way over 10 years as a data scientist so you don't have to!
Ben was also kind enough to give us a 35% off promo code for his book! Use the link: http://mng.bz/n2P5
//Abstract
A great deal of time is spent building out the most effectively tuned model, production-hardened code, and elegant implementation for a business problem. Shipping our precious and clever gems to production is not the end of the solution lifecycle, though, and many-an-abandoned projects can attest to this. In this talk, we will discuss how to think about model attribution, monitoring of results, and how (and when) to report those results to the business to ensure a long-lived and healthy solution that actually solves the problem you set out to solve.
//Bio
Ben Wilson has worked as a professional data scientist for more than ten years. He currently works as a resident solutions architect at Databricks, where he focuses on machine learning production architecture with companies ranging from 5-person startups to global Fortune 100. Ben is the creator and lead developer of the Databricks Labs AutoML project, a Scala-and Python-based toolkit that simplifies machine learning feature engineering, model tuning, and pipeline-enabled modelling. He's the author of Machine Learning Engineering in Action, a primer on building, maintaining, and extending production ML projects.
//Takeaways
Understanding why attribution and performance monitoring is critical for long-term project success
Borrowing hypothesis testing, stratification for latent confounding variable minimization, and statistical significance estimation from other fields can help to explain the value of your project to a business
Unlike in street racing, drifting is not cool in ML, but it will happen. Being prepared to know when to intervene will help to keep your project running.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Ben on LinkedIn: www.linkedin.com/in/benjamin-wilson-arch/
Timestamps:
[00:00] Introduction to Ben Wilson
[00:11] Ben's background in tech
[03:40] Human aspect of Machine Learning in MLOps
[05:51] MLOps is an organizational problem
[09:27] Fragile Models
[12:36] Fraud Cases
[15:21] Data Monitoring
[18:37] Importance of knowing what to monitor for
[22:00] Monitoring for outliers
[24:16] Staying out of Alert Hell
[29:40] Ground Truth
[31:25] Model vs Data Drift on Ground Truth Unavailability
[34:25] Benefit to monitor system or business level metrics
[38:20] Experiment in the beginning, not at the end
[40:30] Adaptive windowing
[42:22] Bridge the gap
[46:42] What scarred you really bad?
MLOps community meetup #57! Last Wednesday we talked to Josh Tobin, Founder, Stealth-Stage Startup.
// Abstract:
Machine learning is quickly becoming a product engineering discipline. Although several new categories of infrastructure and tools have emerged to help teams turn their models into production systems, doing so is still extremely challenging for most companies. In this talk, we survey the tooling landscape and point out several parts of the machine learning lifecycle that are still underserved. We propose a new category of tool that could help alleviate these challenges and connect the fragmented production ML tooling ecosystem. We conclude by discussing similarities and differences between our proposed system and those of a few top companies.
// Bio:
Josh Tobin is the founder and CEO of a stealth machine learning startup. Previously, Josh worked as a deep learning & robotics researcher at OpenAI and as a management consultant at McKinsey. He is also the creator of Full Stack Deep Learning (fullstackdeeplearning.com), the first course focused on the emerging engineering discipline of production machine learning. Josh did his PhD in Computer Science at UC Berkeley advised by Pieter Abbeel.
// Other Links
https://josh-tobin.com
course.fullstackdeeplearning.com
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Josh on LinkedIn: https://www.linkedin.com/in/josh-tobin-4b3b10a9/
Timestamps:
[00:00] Introduction to Josh Tobin
[01:18] Background of Josh into tech
[08:27] We're you guys behind the Rubik's Cube?
[09:26] Rubik's Cube Project
[09:51] "Research is meant to show you what's possible to solve."
[11:07] "That's one of the things that's started to change and I think the MLOps world is maybe a part of that. What I'm excited about this is that people are focusing on the impact of their models."
[13:18] Insights on Testing
[17:11] Evaluation Store
[18:33] "Production Machine Learning is data-driven products that have predictions in the loop."
[23:40] Analyzing and moving forward
[24:02] "My medium term mindset how machine learning is created is that is there's still gonna be humans involved but humans will be more efficient by tools."
[25:50] Is there a market for this?
[27:40] "The long tale of machine learning use cases is becoming part of every products and service more or less the companies create but it's the same way the software part of the products and services the companies create these days. It's going to create an enormous amount of value."
[30:09] Talents
[32:52] Organizational by-ends and knowledge
[35:16] Tools used for Evaluation Store
39:59] Difference from Monitoring Tool
[42:10] Who is the right person to interact in Evaluation Store?
[50:05] Technical challenges of Apple and Tesla
[53:30] "As Machine Learning use cases are getting more and more complicated, higher and higher dimensional data, bigger and bigger models, larger training sets many companies would need in order to continually improve their systems over time."
Coffee Sessions #32 with D. Sculley of Google, The Godfather Of MLOps.
//Bio
D is currently a director in Google Brain, leading research teams working on robust, responsible, reliable and efficient ML and AI. In his time at Google, D worked on nearly every aspect of machine learning, and have led both product and research teams including those on some of the most challenging business problems.
// Links to D. Sculley's Papers
ML Test Score: https://research.google/pubs/pub46555/
Machine Learning: The high-interest credit card of technical debt
https://research.google/pubs/pub43146/
Google Scholar:
https://scholar.google.com/citations?user=l_O64B8AAAAJ&hl=en
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with D. Sculley on LinkedIn: https://www.linkedin.com/in/d-sculley-90467310/
Timestamps:
[00:00] Introduction to D. Sculley
[00:40] Biggest Papers were written by D for Machine Learning
[02:08] What's changed since you wrote those papers?
[02:56] "No 1, there is an MLOps community."
[04:38] Old best practices
[05:12] "The fact that there are jobs titled MLOps, this is different than it was 5 or 6 years ago."
[06:30] Machine Learning Systems then and now
[07:08] "There wasn't the level of general infrastructure that was looking to offer the large scale integrated solutions."
[07:57] ML Test Score
[11:09] "The Test Score was really written for situations where you don't care about one prediction. You care about millions or billions of predictions per day."
[12:27] "In the end, it's not about the score. It's about the process of asking the questions making sure that each of the important questions that you're asking yourself, you have a good answer to."
[13:04] What else is needed in the Test Score?
[14:36] Stratified testing
[17:05] Counterfactual testing
[18:34] Boundaries
[19:15] Dark ages
[20:27] How do you try in Triage?
[21:10] "Reliability is important. There are no small mistakes. If there are errors, they're going to get spotted and publicised. They're going to impact user's lives. The bar is really high and it's worth the effort to ensure strong reliability."
[23:11] How do you build that interest stress test?
[24:39] "I believe that stress test is going to look like a useful way to encode expert knowledge about domain areas."
[25:37] How do I bring robustness?
[26:47] "Because we don't know how to specify the behaviour in advance, testing the behaviour that we wanted to have is a fundamentally hard problem."
[27:22] Underspecification Paper
[30:58] "It's important to be evaluating models on this auto domain stress test and make sure that we understand the implications of what we're thinking about while we are in deployment land."
[32:27] Principal challenges in productionizing Machine Learning
[34:57] "As we expose our models to more specifics, this means there are more potential places our models might be exhibiting unexpected or undesirable behaviour."
[42:37] Splintering of ML Engineering
[46:00] Communities shaping the MLOps sphere
[46:42] "It's much better to have one large community than three smaller communities because of those edufacts."
[47:47] Concept of technical debt in machine learning.
[49:28] "The good idea is to tend to make their way into the community if they are in a form that people can digest and share."
MLOps community meetup #56! Last Wednesday we talked to Daniel Stahl, Head of Data and Analytic Platforms, Regions Bank.
// Abstract:
The Data Science practice has evolved significantly at Regions, with a corresponding need to scale and operationalize machine learning models. Additionally, highly regulated industries such as finance require a heightened focus on reproducibility, documentation, and model controls. In this session with Daniel Stahl, we will discuss how the Regions team designed and scaled their data science platform using DevOps and MLOps practices. This has allowed Regions to meet the increased demand for machine learning while embedding controls throughout the model lifecycle. In the 2 years since the data science platform has been onboarded, 100% of data products have been successfully operationalized.
// Bio:
Daniel Stahl leads the ML platform team at Regions Bank and is responsible for tooling, data engineering, and process development to make operationalizing models easy, safe, and compliant for Data Scientists.
Daniel has spent his career in financial services and has developed novel methods for computing tail risk in both credit risk and operational risk, resulting in peer-reviewed publications in the Journal of Credit Risk and the Journal of Operational Risk. Daniel has a Masters in Mathematical Finance from the University of North Carolina Charlotte.
Daniel lives in Birmingham, Alabama with his wife and two daughters.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Dan on LinkedIn: https://www.linkedin.com/in/daniel-stahl-6685a52a/
Timestamps:
[00:00] Introduction to Ben Wilson
[00:11] Ben's background in tech
[01:17] "How do you do what I have always done pretty well which is being as lazy as possible in order to automate things that I hate doing. So I learned about Regression Problems."
[03:40] Human aspect of Machine Learning in MLOps
[05:51] MLOps is an organizational problem
[09:27] Fragile Models
[12:36] Fraud Cases
[15:21] Data Monitoring
[18:37] Importance of knowing what to monitor for
[22:00] Monitoring for outliers
[24:16] Staying out of Alert Hell
[29:40] Ground Truth
[31:25] Model vs Data Drift on Ground Truth Unavailability
[34:25] Benefit to monitor system or business level metrics
[38:20] Experiment in the beginning, not at the end
[40:30] Adaptive windowing
[42:22] Bridge the gap
[46:42] What scarred you really bad?
MLOps community meetup #55! Last Wednesday we talked to Igor Lushchyk, Data Engineer, Adyen.
// Abstract:
Building Data Science and Machine Learning platforms at a scale-up. Having the main difficulty in finding correct processes and basically being a toddler who learns how to walk on a steep staircase. The transition from homegrown platform to open source solutions, supporting old solutions and maturing them with making data scientists happy.
// Bio:
Igor is a software engineer with more than 10 years of experience. With a background in bioinformatics, he even started PhD but didn't finish it.
As a data engineer, Igor has been working for the last 6 or 7 years, or maybe more - because he was doing almost the same data engineering stuff but his position was named differently.
Igor has been doing a lot of MLOps in 4-5 years now. He doesn't know what he was doing more then - Data Engineering or MLOps. And that’s how this topic came about.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Igor on LinkedIn: https://www.linkedin.com/in/igor-lushchyk/
Timestamps:
[00:00] Introduction to Igor Lushchyk
[02:05] Igor's background in tech
[07:42] Tips you can pass on
[11:05] How these tools work and how they play together and what is underneath?
[13:18] Dedicated MLOps team
[13:55] Central Data Infrastructure Section
[16:57] Transfer over to open-source
[20:24] If you don't plan for production from the beginning, then it's going to be painful trying to go from POC to production.
[22:08] Ho do you handle data lineage?
[25:09] You chose that back in the day but you're regretting it.
[26:34] "Try to use tools which solve 80% of your use cases and maybe 20% you'll have the suffering but at least it's not 100% suffering."
[27:27] Friction points
[28:53] Interaction with Data Scientists
[29:21] "We have alignment sessions. We have different levels of representations. We share our progress."
[32:42] Build verse by decisions
[34:04] When to build or grab an open-source tool
[35:51] Build your own or buy open-source?
[37:11] Certain maturity and a certain number of engineers
[38:11] Startup to go with open-source
[40:14] Correct transition process
[40:56] "There are no other ways but to communicate with data scientists. Your team needs to have a close loop for future priorities, what to take with you and what to leave behind."
[44:51] What to use in monitoring piece
[45:36] Prometheus and Grafana
[48:07] Do you automatic retriggering monitoring of Models set up?
[51:55] Hardware for on Prim model training
[52:38] "Machine Learning model prediction is a spear bomb."
[53:55] War or horror stories
[54:15] "Guys, don't do context switching!"
[55:54] "I won't say that Adyen is a company that allows you to make mistakes but you can make mistakes."
MLOps community meetup #54! Last Wednesday we talked to Laszlo Sragner, Founder, Hypergolic.
// Abstract:
How my experience in quant finance and software engineering influenced how we ran ML at a London Fintech Startup. How to solve business problems with incremental ML? What's the difference between academic and industrial ML?
// Bio:
Laszlo worked as a quant researcher at multiple investment managers and as a DS at the world's largest mobile gaming company. As Head of Data Science at Arkera, he drove the company's data strategy delivering solutions to Tier 1 investment banks and hedge funds. He currently runs Hypergolic (hypergolic.co.uk) an ML Consulting company helping startups and enterprises bring the maximum out of their data and ML operations.
// Takeaways
Continuous evaluation and monitoring is indistinguishable in a well set up product team. Separation of concerns (SE, ML, DevOps, MLOps) is very important for smooth operation, low friction team coordination/communication is key.
To be able to iterate business features into models you need a modelling framework that can express these which is usually a DL package.
DS-es are well motivated to go more technical because they see the rewards of it. All well run (from DS perspective) startups in my experience do the same.
// Other Links
Free eBook about MLPM: https://machinelearningproductmanual.com/
Lightweight MLOps Python package: https://hypergol.ml/
Blog: laszlo.substack.com
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Laszlo on LinkedIn: https://www.linkedin.com/in/laszlosragner/
Timestamps:
[00:00] Introduction to Laszlo Sranger
[02:15] Laszlo's Background
[09:18] Being a Quant, then influenced, what you were doing with the Investment Banks?
[12:24] Do you think this can be applied in different use cases or specific to what you are doing?
[14:41] Do you have any thoughts of a potentially highly opinionated person?
[16:54] Product management in Machine Learning
[24:59] You have to be at a large company or you have to have a large team? [26:38] What are your thoughts on MLOps products helping with product management for ML? Is it an overreach or scope creep?
[32:00] In the messy world of startups due to the big cost of an MVP for NLP is RegEx which means to user feedbacks it's incorporated by tweaking RegEx?
[33:04] Does the ensemble recent models more than older models? If so, what is the decay rate of weights for older models?
[35:40] Since the iterative management model is generic enough for most ML projects, which component of it can be easily generalized and tools built for version control?
[36:38] Topic Extraction: What type of model do you train for that task?
[52:55] Thoughts on Notebooks
[53:34] "I don't hate notebooks. Let's be clear about that. I put it this way, notebooks are whiteboards. You don't want your whiteboards to be your output because it's a sketch of your solution. You want the purest solution."
This is a deep dive into the most recent MLOps Engineering Labs from the point of view of Team 3.
// Diagram Link:
https://github.com/dmangonakis/mlops-lab-example-yelp
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Laszlo on LinkedIn https://www.linkedin.com/in/laszlosragner/
Connect with Artem on LinkedIn: https://www.linkedin.com/in/artem-yushkovsky/
Connect with Paulo on LinkedIn: https://www.linkedin.com/in/paulo-maia-410874119/
Connect with Dimi on LinkedIn:
MLOps community meetup #53! Last Wednesday we talked to Krishna Gade, CEO & Co-Founder, Fiddler AI.
// Abstract:
Training and deploying ML models have become relatively fast and cheap, but with the rise of ML use cases, more companies and practitioners face the challenge of building “Responsible AI.” One of the barriers they encounter is increasing transparency across the entire AI lifecycle to not only better understand predictions, but also to find problem drivers. In this session with Krishna Gade, we will discuss how to build AI responsibly, share examples from real-world scenarios and AI leaders across industries, and show how Explainable AI is becoming critical to building Responsible AI.
// Bio:
Krishna is the co-founder and CEO of Fiddler, an Explainable AI Monitoring company that helps address problems regarding bias, fairness and transparency in AI. Prior to founding Fiddler, Gade led the team that built Facebook’s explainability feature ‘Why am I seeing this?’. He’s an entrepreneur with a technical background with experience creating scalable platforms and expertise in converting data into intelligence. Having held senior engineering leadership roles at Facebook, Pinterest, Twitter, and Microsoft, he’s seen the effects that bias has on AI and machine learning decision-making processes, and with Fiddler, his goal is to enable enterprises across the globe solve this problem.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Krishna on LinkedIn: https://www.linkedin.com/in/krishnagade/
Timestamps:
[00:00] Thank you Fiddler AI!
[01:04] Introduction to Krishna Gade
[03:19] Krisha's Background
[08:33] Everything was fine when you were doing it behind the scenes. But then when you put it out into the wild, we just lost our "baby." It's no longer under our control.
[08:53] "You want to have the assurance of how the system works. Even if it's working fine or if it's not working fine."
[09:37] What else is Explainability? Can you break that down for us?
[13:58] "Explainability becomes the cornerstone technology to have in place for you to build Responsible AI in production."
[14:48] For those used cases that aren't as high stakes, do you feel it's important? Is it up the foodchain?
[18:47] Can we dig into that used case real fast?
[22:01] If it is a human doing it, there's a lot more room for error? Bias or theories can be introduced and then they don't have a basis in reality?
[23:51] Do you need these subject matter experts or someone who is very advanced to be able to set up what the Explainability tool should be looking for at first is it that plug and play and it will know it latches on to the model?
[29:36] Does Explainable AI also entail Explainable Data. I see the point where Explainability can help with getting the insights about data after the model has been trained but should it be handled perhaps more proactively where you unbias the data before training the model on it?
[32:16] As a data scientist, there are situations when the prediction output is expected to support a business decision taken by senior executives. In that case, when the Explainable model gives out a prediction that doesn't align with the stakeholder's expectations, how should one navigate through this tricky situation?
[43:49] How are denen gram clustering for data explainability?
This is a deep dive into the most recent MLOps Engineering Labs from the point of view of Team 1.
// Diagram Link: https://github.com/mlops-labs-team1/engineering.labs#workflow
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Alexey on LinkedIn: https://www.linkedin.com/in/alexeynaiden/
Connect with John on LinkedIn: https://www.linkedin.com/in/johnsavageireland/
Connect with Michel on LinkedIn: https://www.linkedin.com/in/michel-vasconcelos-8273008/
Connect with Varuna on LinkedIn: https://www.linkedin.com/in/vpjayasiri/
Timestamps
[00:00] Introduction to Engineering Labs Participants
[00:34] What are the Engineering Labs?
[01:05] Credits to Ivan Nardini who organized this episode!
[04:24] John Savage Profile
[05:13] Did you want to learn MLFlow before this?
[05:50] Alexey Naiden Profile
[07:26] Varuna Jayasiri Profile
[08:28] Michel Vasconcelos Profile
[10:07] Do something with Pytorch and MLFlow and then figure out the rest: What did the process look like for you all? What have you created?
[13:39] What did the implementation look like? How you went about structuring and coding it?
[17:03] Did you encounter problems along the way?
[20:26] Can you give us a rough overview of what you designed and then where was the first problem you saw?
[23:08] Was there a lot to catch up with or did you feel it was fine. Can you explain how it was?
[24:12] Talk to us about this tool that you have that John was calling out. What was it called?
[24:41] Is this homegrown? You built this?
[24:51] Did you guys implement this when you went to the engineering labs? [26:03] Can you take us through the pipeline and then the serving and what the overall view of the diagram is?
[37:26] For a pet project it works well, but when you wanna start adding a little bit more on top of it wasn't doing the trick?
[38:13] So you see it coming in it's much less of an integral part, another lego building block that is part of the whole thing?
[40:54] Did you all have trouble with Pytorch or MLFlow?
[42:44] Along with that, what was the prompt you were encountering when you were trying to use Torchserve?
[44:27] What are you thinking would have been better in that case?
[49:05] Feedback on how Engineering Labs went
[50:20] Michel: "Engineering Labs should go on. I would like to be a part of it in the next lab."
[51:52] Varuna: "This gives me a tangible thing to look at at any point in time and learn from it."
[53:00] John: "I feel I have an anchor into the world of MLOps from having done this lab."
[55:52] Alexey: "We're at a checkpoint where there are ways we could take"
[56:01] Terraform piece Michel wrote for reproducibility.
MLOps community meetup #52! Last Wednesday we talked to Luke Feeney and Gavin Mendel-Gleason, TerminusDB.
// Abstract:
A look at the open-source 'Git for Data' landscape with a focus on how the various tools fit into the pipeline. Following that scene-setting, we will delve into how and why TerminusDB builds a revision control database from the ground up.
// Takeaways
- Understanding the 'git for data' offering and landscape
- See how to technically approach a revision control database implementation
- Dream of a better tomorrow
// Bio:
Luke Feeney
Operations Lead, TerminusDB
Luke Feeney is Operations Director at TerminusDB. Prior to joining TerminusDB, Luke worked in the Irish Foreign Ministry for a number of years. He served in Ireland’s Permanent Mission to the UN in New York and the Embassies in South Africa and Greece. He was Ireland’s acting Ambassador to Greece for 2016 and 2017. Luke was also the Head of the Government of Ireland’s Brexit Communications Team and the Government Brexit Spokesperson from 2017 to 2018.
Gavin Mendel-Gleason
Chief Technology Officer, TerminusDB
Dr Gavin Mendel-Gleason is CTO of TerminusDB. He is a former research fellow at Trinity College Dublin in the School of Statistics and Computer Science. His research focuses on databases, logic and verification in software engineering. His work includes contributing to the Seshat global historical databank, an ambitious project to record and analyse patterns in human history. He is the inventor of the Web Object Query Language and the primary architect of TerminusDB. He is interested in improving the best practices of the software development community and a strong believer in formal methods and the use of mathematics and logic as disciplines to increase the quality and robustness of software.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Luke on LinkedIn: https://www.linkedin.com/in/luke-feeney/
Connect with Gavin on LinkedIn: https://www.linkedin.com/in/gavinmendelgleason/
Timestamps:
[00:00] MLOps Announcements
[00:17] Slack Community
[00:59] Luke and Gavin's Presentation Style
[01:34] MLOps Community Twitter, LinkedIn and Youtube
[01:45] Introduction to Luke Feeney and Gavin Mendel-Gleason
[04:35] Luke: You wanted Git for Data?
[05:17] Deep Breath || Is there a Git for Data?
[06:30] What is Git for Data?
[08:55] Four Big Buckets
[28:43] Jupiter Notebook
[30:20] Gavin: Collaboration for Structured Data
[31:28] What about gitdifs with gitlfs?
[31:40] Outline: Motivation, Challenges, Solution
[35:35] Motivation: Why Structured Data?
[36:08] Data is Core
[37:34] Challenges: Data is Still in the Dark Ages
[37:40] Structured or Unstructured, we're doing it wrong
[40:15] Managing Data means Collaborating
[45:09] Discoverability and Schema: Structured data requires a real database - not just GIT.
[46:27] Revision Control
[47:00] Collaboration
[48:38] "Git for data, data is the new oil."
[49:01] Why merging is so difficult?
[49:25] "If you have a schema, you can do much more intelligent things."
[52:36] Machine Learning and Revision Control
MLOps community meetup #51! Last Wednesday we talked to Pamela Jasper, AI Ethicist, Founder, Jasper Consulting Inc.
// Abstract:
One of the challenges to the widespread adoption of AI Ethics is not only its integration with MLOps, but the added processes to embed ethical principles will slow and impede Innovation. I will discuss ways in which DS and ML teams can adopt Agile practices for Responsible AI.
// Bio:
Pamela M. Jasper, PMP is a global financial services technology leader with over 30 years of experience developing front-office capital markets trading and quantitative risk management systems for investment banks and exchanges in NY, Tokyo, London, and Frankfurt. Pamela developed a proprietary Credit Derivative trading system for Deutsche Bank and a quantitative market risk VaR system for Nomura. Pamela is the CEO of Jasper Consulting Inc, a consulting firm through which she provides advisory and audit services for AI Ethics governance. Based on her experience as a software developer, auditor and model risk program manager, Pamela created an AI Ethics governance framework called FAIR – Framework for AI Risk which was presented at the NeurIPS 2020 AI conference. Pamela is available as an Advisor, Auditor and Keynote Speaker on AI Ethics Governance. She is a member of BlackInAI, The Professional Risk Managers Industry Association, Global Association of Risk Managers and ForHumanity.
//Takeaways
Agile methods of adopting AI Ethical processes.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Pamela on LinkedIn: https://www.linkedin.com/in/pamela-michelle-j-a5a3a914/
Timestamps:
[00:00] Introduction to Pamela Jasper
[00:17] Pamela's Background
[05:45] Agile IA/Agile Machine Learning, If they are the right fit for each other?
[07:50] What is agile? Not necessarily in and of itself a hard-coded framework.
[08:05] Agile itself based on May 2001 Manifesto is simply a set of values and principles and teams that make decisions around these values and principles.
[10:17] Proposal of Pamela: Let's do Agile with the underlying Ethics that are involved in the ways that you're creating this machine learning. Is that correct?
[10:28] "What I'm suggesting is that Ethics become baked into almost to the mindset of a machine learning engineer, data scientists and in the machine learning operational process for MLOps."
[14:37] "Not all models are created equal"
[15:59] How would be in an Agile way put into practice in your mind?
[36:38] What are the things that would help bridge the gap between AI Ethics and the Agile?
[41:01] It's not that you're trying to bring on the Agile framework to the different pieces of Ethics. It's that you're bringing that into the Agile framework?
[41:21] "We're weaving Ethics into the bedrock of existing Machine Learning practices."
[45:13] How can you really get a diverse team if you're not hiring someone who's there as a diverse person?
[48:59] What would Epics look like if you're baking Ethics?
[52:52] How do you apply Ethics to an ethically questionable domain like gambling?
[54:42] "I think that we can create an AI app for gambling is legal that becomes legal in that construct."
[56:23] Do you think it's possible/desirable to automate any of the ethical considerations in this way?
Coffee Sessions #29 with Jet Basrawi of Satalia, Culture and Architecture in MLOps.
//Bio
Jet started his career in technology as a game designer but became interested in programming. He found he loved it. It was endlessly challenging and deeply enjoyable "Flow" activity. It was also nice to be in demand and earn a living.
In the last several years, Jet been passionate about DevOps as a key strategic practice. About a year ago, he came into the AI world and it is a great place to be for someone like him. The challenges of MLOps and all the things surrounding AI delivery is a great space to work in.
At about the time Jet got into AI the MLops community began, and it was a great experience to come on the journey with Demetrios who was uncovering topics in parallel to him. It was uncanny that each week Demetrios would run a meetup that dealt with exactly the topics he has been trying to reason about.
Jet is very interested in culture and architecture and looking forward to exploring this subject in conversation.
//Takeaways
Insight into the role of culture and architecture in MLOps.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Jet on LinkedIn: https://www.linkedin.com/in/jet-basrawi-4b9ab43/
Timestamps:
[00:00] Introduction to Jet Basrawi
[01:24] Jet's take on MLOps
[02:00] "MLOps - the real Kung fu's in the future" Jet
[02:35] Jet's different opinion on "Tooling is the biggest piece in MLOps".
[04:23] MLOps is a way of life. It's a lifestyle. It's not just tooling.
[04:38] Why do you have to move over to the cultural side and where feel things fail culturally when it comes to machine learning?
[05:47] What you refer to as an orthodox perspective on DevOps and how that place out in your perspective on MLOps?
[06:37] Why do you believe that the separate terminology is coming about and do you believe that this is ultimately harmful to organizations to have this confusion or do you think things should be simplified?
[09:05] As soon as you go down and you're not looking at the big picture. You go down one level and they divert completely, is that your thought too?
[12:30] How do you go about educating yourself and then figuring out how to articulate MLOps or constitutes in your organization?
[16:16] How to do things differently? What are some of your preferred tactics? How to encourage culture change?
[19:02] "Management is NOT Leadership"
[20:13] Why are people stuck in their agile approach?
[23:57] Someone's trying to pick something up for the 1st time and then put it into production, how dangerous that can be?
[25:53] Accepting failure
[29:11] What are some of your principles that helped you communicate to the developers?
[35:33] "It has to dumb down."
[37:43] Annotation [39:37] "Patterntastic"
[41:24] "MLOps is a people problem."
[43:50] Sprint are adequate for machine learning?
[47:03] "Software development is a social activity"
[48:03] "We are all juniors in this field."
//Show Notes
https://www.youtube.com/watch?v=J1WpAJRt3rg Charlie You
https://youtu.be/J36xHc05z-M Manoj https://www.youtube.com/watch?v=vH7UFZZdja8&t=5s Lak design patterns https://www.youtube.com/watch?v=9g4deV1uNZo&t=1s flavios talk
https://continuousdelivery.com/implementing/culture/ westrum culture
https://www.youtube.com/watch?v=Y4H8dW7Ium8&feature=youtu.be&t=109 Jez Humble
MLOps community meetup #50! Last Wednesday we talked to Michael Del Balso, Willem Pienaar and David Aronchick,
// Abstract:
The MLOps tooling landscape is confusing. There’s a complicated patchwork of products and open-source software that each cover some subset of the infrastructure requirements to get ML to production. In this session - we’ll focus on the two most important platforms: model management platforms and feature stores. Model management platforms such as Kubeflow help you get models to production quickly and reliably. Feature stores help you easily build, use, and deploy features. Together, they cover requirements to get models and data to production - the two most important components of any ML project.
In this panel discussion, we’ll be joined by David Aronchick (Co-Founder of Kubeflow), Mike Del Balso (Co-Founder of Tecton) and Willem Pienaar (Creator of Feast). These experts will share their perspective on the challenges of Operational ML and how to build the ideal infrastructure stack for MLOps. They’ll talk about the importance of managing models and data with the same engineering efficiency and best practices that we’ve been applying to application code. They’ll discuss the role of Kubeflow, Feast and Tecton, and share their views on the future of MLOps tooling.
// Bio:
Michael Del Balso
CEO & Co-founder, Tecton
Mike is the co-founder of Tecton, where he is focused on building next-generation data infrastructure for Operational ML. Before Tecton, Mike was the PM lead for the Uber Michelangelo ML platform. He was also a product manager at Google where he managed the core ML systems that power Google’s Search Ads business. Previous to that, he worked on Google Maps. He holds a BSc in Electrical and Computer Engineering summa cum laude from the University of Toronto.
Willem Pienaar
Co-creator, Feast
Willem is currently a tech lead at Tecton where he leads the development of Feast, an open-source feature store for machine learning. Previously he led the ML platform team at Gojek, the Southeast Asian decacorn, which supports a wide variety of models and handles over 100 million orders every month. His main focus areas are building data and ML platforms, allowing organizations to scale machine learning and drive decision making. In a previous life, Willem founded and sold a networking startup.
David Aronchick
Program Manager, Azure Innovations
David leads works in the Azure Innovation Office on Machine Learning. This means he spends most of my time helping humans to convince machines to be smarter. He is only moderately successful at this.
Previously, he led product management for Kubernetes on behalf of Google, launched Google Kubernetes Engine, and co-founded the Kubeflow project. He has also worked at Microsoft, Amazon, and Chef and co-founded three startups.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Michael on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/
Connect with Willem on LinkedIn: https://www.linkedin.com/in/michaeldelbalso/
Connect with David on LinkedIn: https://www.linkedin.com/in/aronchick/
MLOps community meetup #49! Last Wednesday we talked to Lak Lakshmanan, Data Analytics and AI Solutions, Google Cloud.
// Abstract:
Design patterns are formalized best practices to solve common problems when designing a software system. As machine learning moves from being a research discipline to a software one, it is useful to catalogue tried-and-proven methods to help engineers tackle frequently occurring problems that crop up during the ML process. In this talk, I will cover five patterns (Workflow Pipelines, Transform, Multimodal Input, Feature Store, Cascade) that are useful in the context of adding flexibility, resilience and reproducibility to ML in production. For data scientists and ML engineers, these patterns provide a way to apply hard-won knowledge from hundreds of ML experts to your own projects.
Anyone designing infrastructure for machine learning will have to be able to provide easy ways for the data engineers, data scientists, and ML engineers to implement these, and other, design patterns.
// Bio:
Lak is the Director for Data Analytics and AI Solutions on Google Cloud. His team builds software solutions for business problems using Google Cloud's data analytics and machine learning products. He founded Google's Advanced Solutions Lab ML Immersion program and is the author of three O'Reilly books and several Coursera courses. Before Google, Lak was a Director of Data Science at Climate Corporation and a Research Scientist at NOAA.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Lak on LinkedIn: https://www.linkedin.com/in/valliappalakshmanan/
Timestamps:
[00:00] TWIML Con Debate announcement to be hosted by Demetrios on Friday
[00:19] Should data scientists know about Kubernetes? Is it just one machine learning tool to rule them all? Or is it going to be the "best-in-class" tool?
[00:35] Strong opinion of Lak about "Should data scientists know about Kubernetes?"
[05:50] Lak's background into tech
[08:07] Which ones you wrote in the book? Is the airport scenario yours?
[09:25] Did you write ML Maturity Level from Google?
[12:34] How do you know when to bring on perplexity for the sake of making things easier?
[16:06] What are some of the best practices that you've seen being used in tooling?
[20:09] How did you come up with writing the book?
[20:59] How did we decide that these are the patterns that we need to put in the book?
[24:14] Why did I get the "audacity" to think that this is something that is worth doing?
[31:29] What would be in your mind some of the hierarchy of design patterns?
[38:05] Are there patterns out there that are yet to be discovered? How do you balance the exploitable vs the explorable ml patterns?
[42:08] ModelOps vs MLOps
[43:08] Do you feel that a DevOps engineer is better suited to make the transition into becoming a Machine Learning engineer?
[46:07] Fundamental Machine Design Patterns vs Software Development Design Patterns
[49:23] When you're working with the companies at Google, did you give them a toolchain and a better infrastructure or was there more to it? Did they have to rethink their corporate culture because DevOps is often mistaken as just a pure toolchain?
Coffee Sessions #28 with Charlie You of Workday, Lessons learned from hosting the Machine Learning Engineered podcast
//Bio
Charlie You is a Machine Learning Engineer at Workday and the host of ML Engineered, a long-form interview podcast aiming to help listeners bring AI out of the lab and into products that people love. He holds a B.S. in Computer Science from Rensselaer Polytechnic Institute and previously worked for AWS AI.
Charlie is currently working as a Machine Learning Engineer at Workday. He hosts the ML Engineered podcast, learning from the best practitioners in the world.
Check Charlie's podcast and website here:
mlengineered.com
https://cyou.ai/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Charlie on LinkedIn: https://linkedin.com/in/charlieyou/
Timestamps:
[00:00] Introduction to Charlie You
[01:50] Charlie's background on Machine Learning and inspiration to create a podcast
[06:20] What's your experience been so far as the machine learning engineer and trying to put models into production and trying to get things out that has business value?
[07:08] "I started the podcast because as I started working, I had the tingling that machine learning engineering is harder than most people thought, and like way harder than I personally thought."
[08:20] What's an example of that where you target someone in your podcast, you keep that learning and you want an extra meeting the next day and say "Hey, actually I'm starting one of the world's experts on this topics and this is what they said"?
[10:06] In a world of tons of traditional software engineering assets and the process you put in place, how have they adopted what they're doing to the machine learning realm?
[19:00] About your podcast, what are some 2-3 most consistent trends that you've been seeing?
[21:08] Instead of splintering so much as machine learning monitoring infrastructure specialist, are you going to departmentalize it in the future?
[27:22] Is there such a thing as an MLOps engineer right now?
[28:50] "We haven't seen a very vocal, very opinionated project manager in machine learning yet." - Todd Underwood
[30:18] "Similarly with tooling, we haven't seen the emergence of the tools that encode those best practices." Charlie
[31:42] "The day that you don't have to be a subject matter expert in machine learning to feel confident and deploy machine learning products, is the day that you will see the real product leadership in machine learning." Vishnu
[34:12] I'd love to hear your take on some more trends that you've been seeing (Security and Ethics)
[34:41] "Data Privacy and Security is always at the top of any consideration for infrastructure." Charlie
[35:44] That's driven by legal requirements? How do you solve this problem?
[37:27] How do we make sure that if that blows up, you're not left with nothing?
[42:28] In your conversations, have you seen people who goes with cloud provider?
[43:25] Enterprises have much different incentives than startups do.
[45:48] What are some used cases where companies are needing to service their entire needs?
[45:48] What are some used cases where companies are needing to service their entire needs?
[49:18] What are some takeaways that you had in terms of how you think about your career, what experiences you want to build as this MLOps based engineering is moving so fast?
[56:08] "Your edge is never in the algorithm"
Coffee Sessions #27 with Noah Gift of Pragmatic AI Labs, Practical MLOps
// A “Gift” from Above
This week, Demetrios and Vishnu got to spend time with inimitable Noah Gift. Noah is a data science educator, who teaches at Duke, Northwestern, and many other universities, as well as a technical leader through his company Pragmatic AI Labs and past companies.
His bio alone would take up this section of the newsletter, so we invite you to check it out here, as well as the rest of his educational content. Read on for some of our takeaways.
// HOW is as important as WHAT
In our conversation, Noah eloquently pointed out the numerous challenges of bringing ML into production, and especially for making sure it's used positively. It’s not enough to train great models; it’s important to make sure they impact the world positively as their productionized. How models are used is as important as what the model is.
Noah specifically commented on externalities and how’s it incumbent on all MLOps practitioners to understand the externalities created by their models.
// Just get certified
As an educator, Noah has seen front and center how deficits in ML/DS education at the university level have led to the “cowboy” data scientist that doesn’t fit into an effective technical organizational structure. In his courses, Noah emphasizes getting started with off the shelf models and understanding how existing software systems are engineered before committing to building ML systems.
Furthermore, Noah suggested getting certifications as a useful way of upskilling for anyone looking to increase their knowledge base in MLOps, especially by cloud providers.
// Tech Stack Risk
Finally, as many of you do, we debated the relative merits of the major cloud providers (AWS, Azure, and GCP) with Noah. With his vast experience, Noah made a great point about how adopting extremely new tools can sometimes go wrong. In the past, Noah adopted Erlang as a language used in the development of a product. However, as the language never quite took off (in his experience), it became a struggle to hire the right talent to get things done.
Readers, as you go about designing and building the MLOps stack, does any part of the process sound like Noah’s experience with Erlang? Tools or frameworks where downstream adoption may end up fractured? We’d love to hear more!
Definitely check out Noah’s podcast with us for more awesome nuggets on MLOps. Thanks to Noah for taking the time!
https://noahgift.com/
Noah Gift
Machine Learning, Data Science, Cloud & AI Lecturer
His most recent books are:
Pragmatic A.I.: An introduction to Cloud-Based Machine Learning (Pearson, 2018)
Python for DevOps (O’Reilly, 2020).
Cloud Computing for Data Analysis, 2020
Practical MLOps (O'Reilly, 2021 est.)
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Noah on LinkedIn: https://www.linkedin.com/in/noahgift/
[00:00] Introduction to Noah Gift
[03:28] How we can stay pragmatic when it comes to MLOps?
[32:45] The worst excuse that you can give somebody is that "I just do this stuff that's hard, intellectually, but departed, makes it work. That's your job."
[33:34] "In academics, we don't do vocational training, we just teach you theory." "In the Master's Degree, we don't do anything that gets you a job."
[46:33] MLOps vs Cloud Provider
[51:35] GO vs Erlang
MLOps community meetup #48! Last Wednesday, we talked to Manoj Agarwal, Software Architect at Salesforce.
// Abstract:
Serving machine learning models is a scalability challenge at many companies. Most applications require a small number of machine learning models (often < 100) to serve predictions. On the other hand, cloud platforms that support model serving, though they support hundreds of thousands of models, provision separate hardware for different customers. Salesforce has a unique challenge that only very few companies deal with; Salesforce needs to run hundreds of thousands of models sharing the underlying infrastructure for multiple tenants for cost-effectiveness.
// Takeaways:
This talk explains Salesforce hosts hundreds of thousands of models on a multi-tenant infrastructure to support low-latency predictions.
// Bio:
Manoj Agarwal is a Software Architect in the Einstein Platform team at Salesforce. Salesforce Einstein was released back in 2016, integrated with all the major Salesforce clouds. Fast forward to today and Einstein is delivering 80+ billion predictions across Sales, Service, Marketing & Commerce Clouds per day.
//Relevant Links
https://engineering.salesforce.com/flow-scheduling-for-the-einstein-ml-platform-b11ec4f74f97
https://engineering.salesforce.com/ml-lake-building-salesforces-data-platform-for-machine-learning-228c30e21f16
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Manoj on LinkedIn: https://www.linkedin.com/in/agarwalmk/
Timestamps:
[00:00] Happy birthday Manoj!
[00:41] Salesforce blog post about Einstein and ML Infrastructure
[02:55] Intro to Serving Large Number of Models with Low Latency
[03:34] Manoj' background
[04:22] Machine Learning Engineering: 99% engineering + 1% machine learning - Alexey Gregorev on Twitter
[04:37] Salesforce Einstein
[06:42] Machine Learning: Big Picture
[07:05] Feature Engineering [07:30] Model Training
[08:53] Model Serving Requirements
[13:01] Do you standardize on how models are packaged in order to be served and if so, what standards Salesforce require and enforce from model packaging?
[14:29] Support Multiple Frameworks
[16:16] Is it easy to just throw a software library in there?
[27:06] Along with that metadata, can you breakdown how that goes?
[28:27] Low Latency
[32:30] Model Sharding with Replication
[33:58] What would you do to speed up transformation code run before scoring?
[35:55] Model Serving Scaling
[37:06] Noisy Neighbor: Shuffle Sharding
[39:29] If all the Salesforce Models can be categorized into different model type, based on what they provide, what would be some of the big categories be and what's the biggest?
[46:27] Retraining of the Model: Does that deal with your team or is that distributed out and your team deals mainly this kind of engineering and then another team deal with more machine learning concepts of it?
[50:13] How do you ensure different models created by different teams for data scientists expose the same data in order to be analyzed?
[52:08] Are you using Kubernetes or is it another registration engine? [53:03] How is it ensured that different models expose the same information?
**Private data, Data Science friendly**
Data Scientists are always eager to get their hands on more data, in particular, if that data has any value that can be extracted. Nevertheless, in real-world situations, data does not exist in the abundance that we thought existed, in other situations, the data might exist, but not possible to share it with different entities due to privacy concerns, which makes the work of data scientists not only hard, but sometimes even impossible.
// Abstract:
In the last episode of this series, we've decided to bring not one, but two guests to tells us how Synthetic data can unlock the use of data for Data Science teams whenever privacy concerns are a reality. Jean-François Rajotte, Researcher and Resident data Scientist at the University of Columbia and Sumit Mukherjee, Senior Applied Scientist at Microsoft's AI for Good, bring us into more detail their expertise not only, in Synthetic data generation, but in it's mind blowing combination with Federated Learning to take the healthcare sector into the next level of AI adoption.
//Other links to check on Jean-François Rajotte:
https://venturebeat.com/2021/01/20/microsofts-felicia-taps-ai-to-enable-health-providers-to-share-data-anonymously/
https://leap-project.github.io/
//Other links to check on Sumit Mukherjee:
www.sumitmukherjee.com (Sumit research)
https://arxiv.org/abs/2101.07235
https://arxiv.org/pdf/2009.05683.pdf
https://github.com/microsoft/privGAN (PrivGan)
//Final thoughts
Feel free to drop some questions into our slack channel (https://go.mlops.community/slack)
Watch some of the other podcast episodes and old meetups on the channel: https://www.youtube.com/channel/UCG6qpjVnBTTT8wLGBygANOQ
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Fabiana on LinkedIn: https://www.linkedin.com/in/fabiana-clemente/
Connect with Jean-François on LinkedIn: https://www.linkedin.com/in/jfraj/
Connect with Sumit on LinkedIn: https://www.linkedin.com/in/sumitmukherjee2/
Coffee Sessions #26 with Vishnu Rachakonda of Tesseract Health, Daniel Galinkin of iFood, Matias Dominguez of Rappi & Simarpal Khaira of Intuit, Feature Store Master Class.
//Bio
Vishnu Rachakonda
Machine Learning Engineer at Tesseract Health. Coffee sessions co-host but this time his role is one of the all-stars guest speakers.
Daniel Galinkin
One of the co-founders of Hekima, one of the first companies in Brazil to work with big data and data science, with over 10 years of experience in the field. At Hekima, Daniel was amongst the people responsible for dealing with infrastructure and scalability challenges. After iFood acquired Hekima, he became the ML Platform Tech Lead for iFood.
Matias Dominguez
A 29-year-old living in Buenos Aires, past 4.5 years working on fraud prevention. Previously at MercadoLibre and other random smaller consulting shops.
Simarpal Khaira
Simarpal is the product manager driving product strategy for Feature Management and Machine Learning tools at Intuit. Prior to Intuit, he was at Ayasdi, a machine learning startup, leading product efforts for machine learning solutions in the financial services space. Before that, he worked at Adobe as a product manager for Audience Manager, a data management platform for digital marketing.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community:
https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup:
https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Daniel on LinkedIn: https://www.linkedin.com/in/danielgalinkin/
Connect with Matias on LinkedIn: https://www.linkedin.com/in/mndominguez/
Connect with Simarpal on LinkedIn: https://www.linkedin.com/in/simarpal-khaira-6318959/
Timestamps:
[00:00] Introduction to guest speakers.
[00:33] Vishnu Rachakonda Background
[01:00] Guest speakers' Background
[03:13] Are Feature Stores for everyone?
[04:02] Guest speakers' Feature Store background
[17:09] How do you go about gathering requirements for a Feature Store and customize it?
[17:34] Guest speakers' process for Feature Store
[31:14] What solution are we actually trying to build?
[36:42] How do you ensure consistency in your transformation logic and in your process generating features?
[43:39] In terms of versioning that transformation logic and knowledge that goes into creating Feature Stores and allowing them to be reusable and consistent, how are you going to grapple with that?
[48:06] How do you bake in best practices into the services that you offer?
[49:34] "It's too possible for you to do something wrong. You have to specify that wrong thing. That makes it harder to do that wrong thing." Daniel
[51:54] "It starts with changing the mindset. Making people getting the habit of what is the value here. Then you are producing features for consumers because tomorrow you could become a consumer. Write it in a way as you want to consume somebody's feature." Simar
[56:51] "As part of that process, it should come with everyone's best practices to actually improve all features" Matias
MLOps community meetup #47! Last Wednesday, we talked to Adrià Romero, Founder and Instructor at ProductizeML.
// Abstract:
In this talk, we tackled:
- Motivations and mission behind ProductizeML.
- Common friction points and miscommunication between technical and management/product teams, and how to bridge these gaps.
- How to define ML product roadmaps, (and more importantly, how to get it signed off by all your team).
- Best practices when managing the end-to-end ML lifecycle.
/ Takeaways:
- Self-study guide that reviews the end-to-end ML lifecycle starting with some ML theory, data access and management, MLOps, and how to wrap up all these pieces in a viable but still lovable product.
- Free and collaborative self-study guide built by professionals with experience on different stages from the ML lifecycle.
// Bio:
Adrià is an AI, ML, and product enthusiast with more than 4 years of professional experience on his mission to empower society with data and AI-driven solutions.
Born and raised in the beautiful and sunny Barcelona, he began his journey in the AI field as an applied researcher at the Florida Atlantic University, where he published some of the first deep learning works in the healthcare sector. Attracted by the idea of deploying these ideas to the real world, he then joined Triage, a healthcare startup building healthcare solutions powered by AI, such a smartphone app able to detect over 500 skin diseases from a picture. During this time, he has given multiple talks at conferences, hospitals, and institutions such as Novartis and Google. Previously, he interned at Huawei, Schneider Electric, and Insight Center for Data Analytics.
Early this year, he started crafting ProductizeML, An Instruction and Interactive Guide for Teams Building Machine Learning Products where he and a team of AI & product specialists carefully prepare content to assist on the end-to-end ML lifecycle.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Adria on LinkedIn: https://www.linkedin.com/in/adriaromero/
References mentioned on this episode:
https://en.wikipedia.org/wiki/ImageNet
https://twitter.com/productizeML https://course.productize.ml/
https://github.com/ProductizeML/gitbook
https://adria756514.typeform.com/to/V4BDqjYA - Newsletter Signup
https://www.buymeacoffee.com/
Timestamps:
[00:00] Introduction to Adrià Romero
[00:32] How did you get into tech?
[02:16] ImagiNet Project (Visual Recognition Challenge)
[06:49] Visual Recognition with Skin Lesions
[07:05] Fundamental vs Applied Research (Academia experience)
[08:44] Motivation for technology
[14:55] Transition to ProductizeML
[19:09] ProductizeML Context
[23:50] What was its that made you think that Education is probably more powerful?
[24:21] ProductizeML Objective
[26:55] Ethics: Do you want to put that in there later?
[30:12] ProductizeML Content Format and Tools
[34:07] ProductizeML Catalogue
[39:28] ProductizeML Audience Target
[42:54] "Buy me a coffee" platform
[48:29] Do you ever foresee with the educational being more vertical-specific?
The revolution of Federated Learning - And we're back with another episode of the podcast When Machine Learning meets Privacy! For the episode #8 we've invited Ramen Dutta, a member of our community and founder of TensoAI.
// Abstract:
In this episode, Ramen explain us the concept behind Federated Learning, all the amazing benefits and it's applications in different industries, particularly in agriculture. It's all about not centralizing the data, sound awkward? Just listen to the episode.
//Other links to check on Ramen:
https://www.linkedin.com/in/tensoai/
//Final thoughts
Feel free to drop some questions into our slack channel (https://go.mlops.community/slack)
Watch some of the other podcast episodes and old meetups on the channel: https://www.youtube.com/channel/UCG6qpjVnBTTT8wLGBygANOQ
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Fabiana on LinkedIn: https://www.linkedin.com/in/fabiana-clemente/
Connect with Ramen on LinkedIn: https://www.linkedin.com/in/tensoai/
Coffee Sessions #25 with Marian Ignev of CloudStrap.io & SashiDo.io, Most Underrated MLOps Topics.
//Bio
Marian a passionate entrepreneur, backend dude & visionary.
These are the three main things described to Marian very well:
Marian's everyday routines include making things happen and motivating people to work hard and learn all the time because I think success is a marathon, not just a sprint!
Marian loves to communicate with bright creative minds who want to change things.
His favorite professional topics are backend stuff, ops, infra, IoT, AI, Startups, Entrepreneurship.
In his free time, he loves to cook for my family.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Marian on LinkedIn: https://www.linkedin.com/in/mignev/
Timestamps:
[00:00] Introduction to Marian Ignev
[01:57] Marian's background
[10:06] Who do you need and what should they be doing?
[18:05] How are you solving problems at your company?
[27:22] What are your thoughts around ML tooling? Why that hasn't happened yet and why it will change?
[33:16] You can't actually figure out what's the main focus of ML tooling services.
[34:14] "Start small and start simple to focus only on a small problem that you can bring more than the others."
[37:14] How are you making the case for how standardization to occur in your initial MLOps?
[38:08] "If you're doing a mistake somewhere, do it everywhere because it will be very easy to find and replace it after that."
[41:50] How do you model monitoring?
[47:19] How would you recommend people to get started?
[49:00] Ecosystem of Machine Learning in Eastern Europe
Other links you can check on Marian:
https://www.sashido.io/
https://www.cloudstrap.io/
https://twitter.com/mignevm.ignev.net/blog/ (Blog)
m.ignev.net/ (Personal Website)
fridaycode.net/ (FridayCode)
MLOps community meetup #46! Last Wednesday, we talked to Hendrik Brackmann, Director of Data Science and Analytics at Tide.
// Abstract:
Tide is a U.K.-based FinTech startup with offices in London, Sofia, and Hyderabad. It is one of the first, and the largest business banking platform in the UK, with over 150,000 SME members. As of 2019, one of Tide’s main focuses is to be data-driven. This resulted in the forming of a Data Science and Analytics Team with Hendrik Brackmann at its head. Let's witness Hendrik's personal anecdotes in this episode!
// Bio:
After studying probability theory at the University of Oxford, Hendrik joined MarketFinance, an SME lender, in order to develop their risk models. Following multiple years of learning, he joined Finiata, a Polish and German lender in order to build out their data science function. Not only did he succeed in improving the risk metrics of the company he also learnt to manage a different department as interim Head of Marketing.
Hendrik's job as Director of Data Science and Analytics at business bank Tide is to oversee data engineering, data science, insights and analytics and data governance functions of Tide.
// Final thoughts
Please feel free to drop some questions you may have beforehand into our slack channel
(https://go.mlops.community/slack)
Watch some old meetups on our youtube channel:
https://www.youtube.com/channel/UCG6qpjVnBTTT8wLGBygANOQ
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Hendrik on LinkedIn: https://www.linkedin.com/in/hendrik-brackmann-b2b5477a/
Timestamps:
[00:00] Introduction to Hendrik Brackmann
[01:54] Hendrik's background into tech
[03:22] First Phase of the three epic journeys of Hendrik
[08:05] Were there some hiccups you were running into as you're trying to make things better?
[10:50] Any other learnings that you got from that job that you want to pass along to us?
[11:50] You were doing all batch at that point, right?
[12:35] Phase 2: of Hendrik's epic journey
[15:11] Did you eventually cut down at the time that it took?
[15:50] Breakdown of Transformation terminologies and its importance
[19:03] What are some things that you would never do again?
[20:32] How did you see things more clearly? [22:30] Phase 3: Moving on to Tide
[24:46] Have you only worked with teams with one programming language?
[30:47] Did you try to open-source solutions or did you just go right out to buy it?
[33:12] What is real-time for you? How much latency is there? How much time do you need?
[37:18] What stage did you realize to get the feature store?
[40:09] What would you recommend from a maturity standpoint to get a feature store?
[41:20] Can you summarize some of the greatest problems that the feature stores solve for you?
[42:22] What problems does a feature store introduces if any?
[44:39] Where do the model and the feature start from the perspective of a system in the engineering?
[49:15] You need a good data management in feature stores
[50:21] Have you ever used or built any feature stores that explicitly handle units and does dimensional analysis on derived features?
[54:46] What kind of models do you have up at the moment and how do you test and monitor and deploy the models?
Coffee Sessions #24 with Sara Robinson of Google, Machine Learning Design Patterns co-hosted by Vishnu Rachakonda.
//Bio
Sara is a Developer Advocate for Google Cloud, focusing on machine learning. She inspires developers and data scientists to integrate ML into their applications through demos, online content, and events. Before Google, she was a Developer Advocate on the Firebase team. Sara has a Bachelor’s degree from Brandeis University. When she’s not writing code, she can be found on a spin bike or eating frosting.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Sara on LinkedIn: https://www.linkedin.com/in/sara-robinson-40377924/
Timestamps:
[00:00] Introduction to Sara Robinson
[01:38] Sara's Background into tech
[04:54] What were some things that jumped out at you right away with Machine Learning that is different?
[07:44] Sara's Transition to the Machine Learning realm.
[08:36] What is the role of a Developer Advocate?
[11:41] Compared to traditional software developer advocacy, what stands out to you as being different, unique, perhaps more fun about working in the Machine Learning realm as a Developer Advocate?
[13:40] "No one person has it right."
[15:27] Given how new this space is, how did you go about writing a book? What leads you to write this book (Machine Learning Design Patterns)? [19:00] Process of deciding to write the book
[21:46] What is it that made the focus of these design patterns?
[25:07] Who's the reader that you think who's gonna have this book on their shelf as a reference?
[26:42] How would you advise readers to go about reconciling this domain-based needs and the design patterns that you may suggest or identify? [31:20] Can you tell us about a time that some of the design patterns as you're learning with your co-authors has been useful to you?
[36:50] Workflow Pipeline breakdown in the book
[42:23] How do you think about that level of maturity in terms of thinking about the design patterns?
[46:06] How do I communicate in design pattern? What if there is resistance to formalization or implementational structure because it might prevent creativity or reiteration?
[49:32] Pre-bill and custom components of Pipeline Frameworks
[51:28] How do we know to do the next step or stay in Feature Store patterns? [56:07] Are we going to see the convergence of tools and frameworks soon?
Resources referenced in this episode:
https://www.oreilly.com/library/view/machine-learning-design/9781098115777/
https://www.amazon.com/Machine-Learning-Design-Patterns-Preparation/dp/1098115783 https://books.google.com.ph/books/about/Machine_Learning_Design_Patterns.html?id=djwDEAAAQBAJ&redir_esc=y
https://amzn.to/38tM22C
https://sararobinson.dev/2020/11/17/writing-a-technical-book.html
Coffee Sessions #23 with Todd Underwood of Google, Followups from OPML Talks on ML Pipeline Reliability co-hosted by Vishnu Rachakonda.
//Bio
Todd is a Director at Google and leads Machine Learning for Site Reliability Engineering Director. He is also Site Lead for Google’s Pittsburgh office. ML SRE teams build and scale internal and external ML services and are critical to almost every Product Area at Google.
Before working at Google, Todd held a variety of roles at Renesys. He was in charge of operations, security, and peering for Renesys’s Internet intelligence services that are now part of Oracle's Cloud service. He also did product work for some early social products that Renesys worked on. Before that Todd was Chief Technology Officer of Oso Grande, an independent Internet service provider (AS2901) in New Mexico.
//Other links referenced by Todd:
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/
Connect with Todd on LinkedIn: https://www.linkedin.com/in/toddunder/
Timestamps:
[00:00] Intro to Todd Underwood
[02:04] Todd's background
[08:54] What's kind of vision do you "paint"?
[14:54] Playing a little bit "devil's advocate." Do you think that's even possible?
[19:36] "Start serving to make sure of having the possibility to get it out." How do you feel about that?
[23:56] What advise could you give to other people who wanted to bring in ML professionals into their companies to make ML useful for them? [29:53] Is it useful to use these new models?
[32:25] Do you feel like there would be a point where there would be a standard procedure?
[35:50] How machine learning breaks
[40:44] As an engineering leader, what's your advice to other engineering leaders in terms of how to make that reflection on your team needs and failures...?
[48:42] It's the design that you're looking at as the problem, not the person.
[56:27] Do we think that people sold a bunch of stuff and now we were left with the results?
[1:00:46] Recommendations on readings, things to do to better hone our craft.
[1:03:35] The more you explore, the more you realize, what's going on? Where can I learn from?
[1:05:00] Since you are in the mode of predicting things and philosophical background, where are you seeing the industry going in the next 5 years as we create it?
Resources referenced in this episode:
https://www.youtube.com/watch?v=Nl6AmAL3i08&feature=emb_title&ab_channel=USENIX
https://www.youtube.com/watch?v=hBMHohkRgAA&ab_channel=USENIX
https://youtu.be/0sAyemr6lzQ https://youtu.be/EyLGKmPAZLY
https://www.usenix.org/conference/opml20/presentation/papasian
https://www.usenix.org/system/files/login/articles/02_underwood.pdf
https://storage.googleapis.com/pub-tools-public-publication-data/pdf/da63c5f4432525bcaedcebeb50a98a9b7791bbd2.pdf
MLOps community meetup #45! Last Wednesday, we talked to Joe Reis, CEO/Co-Founder of Ternary Data.
// Abstract:
The fact is that most companies are barely doing BI, let alone AI. Joe discussed ways for companies to build a solid data foundation so they can succeed with machine learning. This meetup covers the continuum from cloud data warehousing to MLOps.
// Bio:
Joe is a Data Engineer and Architect, Recovering Data Scientist, 20 years in the data game. Joe enjoys helping companies make sense of their culture, processes, and architecture so they can go from dreaming to doing. He’s certified in both AWS and Google Cloud. When not working, you can find Joe at one of the two groups he co-founded—The Utah Data Engineering Meetup and SLC Python. Joe also sits on the board of Utah Python, a non-profit dedicated to advocating Python in Utah.
// Other links to check on Joe:
https://www.youtube.com/channel/UC3H60XHMp6BrUzR5eUZDyZg
https://josephreis.com/
https://www.ternarydata.com/
https://www.linkedin.com/pulse/what-recovering-data-scientist-joe-reis/
https://www.linkedin.com/pulse/should-you-get-tech-certification-depends-joe-reis/
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Joe on LinkedIn: https://www.linkedin.com/in/josephreis/
Timestamps:
[00:23] How did you get into tech? What brought you on to the journey into data?
[04:50] You got into the auto ML and you decided to branch out and do your own thing? How did that happen?
[08:18] What is it with BI and then making that jump to ML?
[11:00] How have you seen Machine Learning fall flat with trying to shoehorn Machine Learning on top of the already weak foundation of BI?
[13:45] Let's imagine we're doing BI fairly well and now we want to jump to Machine Learning. Do we have to go out and reinvent the whole stack or can we shoehorn it on?
[15:36] How do you move from BI to ML?
[18:24] What do you mean by realtime?
[20:35] Managed Services in DevOps
[23:30] The maturity isn't there yet
[26:03] Where would you draw the line between BI and AI?
[30:45] What are the things is Machine Learning an overkill for?
[33:43] Are you thinking about what data sets to collect and how different do those vary?
[35:18] "Software Engineering and Data Engineering are basically going to merge into one."
[38:27] What do you usually recommend moving from BI to AI?
[40:45] What is "strong data foundation" in your eyes?
[42:47] "MLFlow to gateway drug." What's your take on it?
[46:25] In this pandemic, how easy is it for you to pivot to a new provider?
[49:10] Vision of companies starts coming together on different parts of the stack in the Machine Learning tools.
Coffee Sessions #22 with Carl Steinbach of LinkedIn, Deep in the Heart of Data.
//Bio
Carl is a Senior Staff Software Engineer and currently the Tech Lead for LinkedIn's Grid Development Team. He is a contributor to Emerging Architectures for Modern Data Infrastructure
//Other links referenced by Carl:
https://rise.cs.berkeley.edu/wp-content/uploads/2017/03/CIDR17.pdf
https://www.youtube.com/watch?v=-xIai_FvcSk&ab_channel=WePayEngineering
https://softwareengineeringdaily.com/2019/10/23/linkedin-data-platform-with-carl-steinbach/
https://www.slideshare.net/linkedin/carl-steinbach-open-source
https://dreamsongs.com/RiseOfWorseIsBetter.html
https://engineering.linkedin.com/blog/2017/03/a-checkup-with-dr--elephant--one-year-later
https://engineering.linkedin.com/
https://engineering.linkedin.com/blog/2018/11/using-translatable-portable-UDFs
https://a16z.com/2020/10/15/the-emerging-architectures-for-modern-data-infrastructure/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Carl on LinkedIn: https://www.linkedin.com/in/carlsteinbach/
Timestamps:
[00:00] Introduction to Carl Steinbach
[00:44] Carl's background
[04:51] Breakdown of Transpiler
[10:55] Advantages of Decoupling the Execution Layer
[15:25] Differences between UDF (user-defined function) Functions and Views
[18:45] How do you ensure the reproducibility of these Views?
[23:58] Data structure evolution
[27:55] Are Data Lakes and Data Warehouse fundamentally different things or are they on a path towards conversion?
[33:37] It's inevitable that people will start doing machine learning on databases
[36:01] Who gets permission on what, especially when it comes to data and how sensitive things can be?
[41:27] Security aspect of data
[43:40] Does it require a level of obstruction on top of the data of the file system?
[45:48] Why do we go back and go forward which sets this trend?
ML and Encryption - It's all about secure insights #7! In this episode, we've invited Théo Ryffel, Founder of Arkhn and founding member of the Open-Mined community.
// Abstract:
In this episode, Théo introduces us to the concept of encrypted Machine Learning, when and the best practices to have it applied in the development of Machine Learning based solutions, and the challenges of building a community.
//Other links to check on Théo:
https://twitter.com/theoryffel
https://arxiv.org/pdf/1811.04017.pdf
https://arxiv.org/pdf/1905.10214.pdf
//Final thoughts
Feel free to drop some questions into our slack channel (https://go.mlops.community/slack)
Watch some of the other podcast episodes and old meetups on the channel: https://www.youtube.com/channel/UCG6qpjVnBTTT8wLGBygANOQ
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Fabiana on LinkedIn: https://www.linkedin.com/in/fabiana-clemente/
Connect with Théo on LinkedIn: https://www.linkedin.com/in/theo-ryffel
**Privacy-preserving ML with Differential Privacy**
Differential privacy is without a question one of the most innovative concepts that came around in the last decades, with a variety of different applications even when it comes to Machine Learning. Many are organizations already leveraging this technology to access and make sense of their most sensitive data, but what is it? How does it work? And how can we leverage it the most?
To explain this and provide us a brief intro on Differential Privacy, I've invited Christos Dimitrakakis. Professor at University, counts already with multiple publications (more than 1000!!!) in the areas of Machine Learning, Reinforcement Learning, and Privacy.
Useful links:
Christos Dimitrakakis list of publications
Differential privacy for Bayesian inference through posterior sampling
Authors: Christos Dimitrakakis, Blaine Nelson, Zuhe Zhang, Aikaterini Mitrokotsa, Benjamin IP Rubinstein
Differential privacy use cases
Open-source differential privacy projects
Open-source project for Differential Privacy in SQL databases
MLOps community meetup #44! Last Wednesday, we talked to Savin Goyal, Tech lead for the ML Infra team at Netflix.
// Abstract:
In this conversation, Savin talked about some of the challenges encountered and choices made by the Netflix ML Infrastructure team while developing tooling for data scientists.
// Bio:
Savin is an engineer on the ML Infrastructure team at Netflix. He focuses on building generalizable infrastructure to accelerate the impact of data science at Netflix.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Savin on LinkedIn: https://www.linkedin.com/in/savingoyal/
Timestamps:
[00:00] Background of Savin Goyal
[02:41] Breakdown of Metaflow
[05:44] In the stack, where does Metaflow stand?
[13:23] Where does Metaflow start in Runway Project?
[15:27] What tools or storage does Netflix use for DataOps, ie: the front-end management of data sets and how does that integrate with Metaflow? [18:56] Recommender Systems: Can you explain the other areas that you're using Machine Learning?
[22:27] What do you feel is the hardest part of building an operating Machine Learning workflow? [28:45] 3 Pillars: Reproducibility, Scalability, Usability.
[36:05] You give so much power to people. How do you keep them from going overboard?
[37:47] Can you explain this Pillar of Usability?
[41:09] Road-based access control has been coming up a lot recently. Does Metaflow do something specific for that?
[44:49] What are some learnings that come across that you didn't have since you open-sourced when you were working at Netflix?
[48:10] What kind of trends you have been seeing? Where do you feel like the market is going?
[50:33] Have you seen some companies really interested in Metaflow? How have you been seeing them combine other tools that are out there?
Coffee Sessions #21 with Benjamin Rogojan of Seattle Data Guy, A Conversation with Seattle Data Guy
//Bio
Ben has spent his career focused on all forms of data. He has focused on developing algorithms to detect fraud, reduce patient readmission and redesign insurance provider policy to help reduce the overall cost of healthcare. He has also helped develop analytics for marketing and IT operations in order to optimize limited resources such as employees and budget. Ben privately consults on data science and engineering problems both solo as well as with a company called Acheron Analytics. He has experience both working hands-on with technical problems as well as helping leadership teams develop strategies to maximize their data.
//Other links you can check Ben on
https://www.theseattledataguy.com/mlops-vs-aiops-what-is-the-difference/#page-content
https://medium.com/@benrogojan
https://www.kdnuggets.com/2020/01/data-science-interview-study-guide.html
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Ben on LinkedIn: https://www.linkedin.com/in/benjaminrogojan/
Timestamps
[00:00] Intro to Benjamin Rogojan
[01:22] Ben's background
[03:30] What are some of your learnings/key things that jumped out of you?
[08:15] Agile and Data Science
[10:28] Likelihood of failure
[13:05] Sometimes you have to wait
[15:11] Defining your data science process
[19:55] Layer of communication is important between the data scientists and higher-ups
[21:29] How do you navigate challenges? Are there any tools or processes you quantify to work with your clients?
[24:30] How do you show the value of your work using monitoring and observability
[27:58] How can we be better communicators?
[31:15] Have you seen other roles that really helped the jell of the team?
[33:50] What are your interests? What are you passionate about at the moment?
[34:29] Is there something new you're learning at the moment?
[37:55] Do you have a process about how you figure out even data science or ML is right for a company?
[39:33] Do you have a blog about the process you follow?
[41:24] What is one negative wisdom that you want to share with the community?
[44:35] How did you come up with the company name Seattle Data Guy?
Links mentioned in this episode:
https://medium.com/@benrogojan
https://www.cprime.com/resources/blog/agile-methodologies-how-they-fit-into-data-science-processes/
https://www.coriers.com/the-data-science-interview-study-guide/
https://medium.com/@SeattleDataGuy/from-data-scientist-to-data-leader-workshop-c6be69698af
https://towardsdatascience.com/4-must-have-skills-every-data-scientist-should-learn-8ab3f23bc325
Coffee Sessions #20 with Neal Lathia of Monzo Bank, talking about Monzo Bank - An MLOps Case Study
//Bio
Neal is currently the Machine Learning Lead at Monzo in London, where his team focuses on building machine learning systems that optimise the app and help the company scale. Neal's work has always focused on applications that use machine learning - this has taken him from recommender systems to urban computing and travel information systems, digital health monitoring, smartphone sensors, and banking.
//Talk Takeaways
Monzo Bank has a small, but a very impactful team continuously learning new things. Optimistically do their utmost to avoid “throwing problems over the wall,” and so they build systems, iterate on machine learning models, and collaborate very closely with each other and with many folks across the business.
Hopefully, all of that paints a picture of a team that aims to bring real and valuable machine learning systems to life. Monzo does not spend time trying to advance the state-of-the-art in machine learning or tweak models to absolute perfection.
//Other links you can check Neal on
Personal Website: http://nlathia.github.io/
Research: http://nlathia.github.io/research/
Press & Speaking: http://nlathia.github.io/public/
http://nlathia.github.io/2020/06/Customer-service-machine-learning.html
http://nlathia.github.io/2020/10/ML-and-rule-engines.html
http://nlathia.github.io/2020/10/Monzo-ML.html http://nlathia.github.io/2019/09/Large-NLP-in-prod.html http://nlathia.github.io/2020/07/Shadow-mode-deployments.html https://github.com/operatorai
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Neal on LinkedIn: https://www.linkedin.com/in/nlathia/
Timestamps:
[00:00] Intro to Neal Lathia
[02:48] Background of Monzo Bank
[05:06] Problems you're solving with Machine Learning at Monzo?
[08:36] Why do you think it's fairly easy to frame a lot of problems using Machine Learning?
[11:56] How do you decide on rule-based or Machine learning?
[15:33] Team Structure
[19:18] What are some challenges like size, latency and the like?
[21:52] How have you addressed learning skills/challenges in your team?
[26:17] Do you have something that connects your team with all the metadata you have?
[27:14] Are you also having the monitoring models in your dashboard or is that something else?
[28:51] Why should I bring another tool that the company is not familiar with when we already have one?
[31:43] Do you feel like there will be a point in time where you need to buy a tool because one problem is taking so much of your time?
[38:30] Engineering optimization teams for machine learning?
[40:34] Take us through the idea to production?
[46:29] How do you deal with reproducibility?
[49:48] Do you have ethics people on the team?
[54:12] Why are you using GCP and AWS?
[56:09] What are these different used cases and how do they differ?
[57:57] How do you address applications that don't work?
**The intersection between DataOps and privacy**
DataOps is considered by many as the new era of data management, a set of principles that emphasizes communication, collaboration, integration, and automation of cooperation between the different teams in an organization that have to deal with data: data engineers, data scientists to data analysts.
But is there any relation between DataOps and data privacy protection? Can organizations leverage DataOps to ensure that their data is privacy compliant?
For this episode we've invited Lars Albertsson founder of Scling and former Data Engineer at Spotify, Lars has been educating organizations on how to get value from data and engineering efficiency!
You can easily find him and reach out on Twitter and LinkedIn.
Don't forget to join the MLOps.Community if you are not yet a member.
Useful links:
What is DataOps - https://www.ibm.com/blogs/journey-to-ai/2019/12/what-is-dataops/
Data engineering reading list - https://www.scling.com/reading-list/
Data engineering courses - https://www.scling.com/courses/
**Are Privacy Enhancing Technologies a myth**
Data Privacy and machine learning are here to stay, and there’s no doubt they’re the hot trends to be following. But do they need to clash with each other? Can we have these titans to co-exist? It seems like finally 2020 and 2021 will be the years where Privacy Enhancing Technologies. But after all what are they? How are these techs being used and leveraged by organizations?
Useful links:
https://medium.com/@francis_49362/differential-privacy-not-a-complete-disaster-i-guess-d0345a76a5af
Facebook and DIfferential Privacy
Coffee Sessions #19 with Barr Moses of Monte Carlo, Introducing Data Downtime: How to Prevent Broken Data Pipelines with Observability co-hosted by Vishnu Rachakonda
//Bio
Barr Moses is CEO & Co-Founder of Monte Carlo, a data observability company backed by Accel and other top Silicon Valley investors. Previously, she was VP Customer Operations at customer success company Gainsight, where she helped scale the company 10x in revenue and among other functions, built the data/analytics team. Prior to that, she was a management consultant at Bain & Company and a research assistant at the Statistics Department at Stanford. She also served in the Israeli Air Force as a commander of an intelligence data analyst unit. Barr graduated from Stanford with a B.Sc. in Mathematical and Computational Science.
//Talk Takeaways
As companies become increasingly data-driven, the technologies underlying these rich insights have grown more and more nuanced and complex. While our ability to collect, store, aggregate, and visualize this data has largely kept up with the needs of modern data teams (think: domain-oriented data meshes, cloud warehouses, data visualization tools, and data modelling solutions), the mechanics behind data quality and integrity has lagged.
To keep pace with data’s clock speed of innovation, data engineers need to invest not only in the latest modelling and analytics tools but also technologies that can increase data accuracy and prevent broken pipelines. The solution? Data observability, the next frontier of data engineering and a pillar of the emerging Data Reliability category and the fix for eliminating data downtime.
During this talk, listeners will learn about:
//About Monte Carlo As businesses increasingly rely on data to drive better decision making, it’s mission-critical that this data is accurate and reliable. Billed by Forbes as the New Relic for data teams and backed by Accel and GGV, Monte Carlo solves the costly problem of broken data through their fully automated, end-to-end data reliability platform. Data teams spend north of 30% of their time tackling data quality issues, distracting data engineers, data scientists, and data analysts from working on revenue-generating projects. Providing full coverage of your data stack – all the way from data lake and warehouse to analytics dashboard – Monte Carlo’s platform empowers companies such as Eventbrite, Compass, Vimeo, and other enterprises to trust their data, saving time and money and unlocking the potential of data.
//Other links you can check Barr on
Learn more about Monte Carlo: https://www.montecarlodata.com
What is data downtime? https://www.montecarlodata.com/the-rise-of-data-downtime/
What is data observability? https://www.montecarlodata.com/data-observability-the-next-frontier-of-data-engineering/
How data observability prevents broken data pipelines: https://www.montecarlodata.com/data-observability-how-to-prevent-your-data-pipelines-from-breaking/
MLOps community meetup #43! Last Wednesday, we talked to Nathan Benaich, General Partner at Air Street Capital and Timothy Chen, Managing Partner at Essence VC about The MLOps Landscape.
// Abstract:
In this session, we explored the MLOps landscape through the eyes of two accomplished investors. Tim And Nathan shared with us their experience in looking at hundreds of ML and MLOps companies each year to highlight major insights they have gained. What do the ML infrastructure and tooling landscape look like at the moment? Where have they been seeing patterns emerge? What do they expect to see happen within the market in the next couple of years? What current tools out there are the most interesting to them? And last but not least how do they go about selecting which companies to invest in.
// Bio:
Nathan Benaich is the Founder and General Partner of Air Street Capital, a venture capital firm investing in early-stage AI-first technology and life science companies. The team’s investments include Mapillary (Acq. Facebook), Graphcore, Thought Machine, Tractable, and LabGenius. Nathan is Managing Trustee of The RAAIS Foundation, a non-profit with a mission to advance education and open-source research in common good AI. This includes running the annual RAAIS summit and funding fellowships at OpenMined. Nathan is also co-author of the annual State of AI Report. He holds a PhD in cancer biology from the University of Cambridge and a BA from Williams College.
Timothy Chen is the Managing Partner at Essence VC, with a decade of experience leading engineering in enterprise infra and open source communities/companies.
Prior to Essence, Tim was the SVP of Engineering at Cosmos, a popular open-source blockchain SDK. Prior to Cosmos, Tim cofounded Hyperpilot with Stanford Professor Christos Kozyrakis which later exited to Cloudera. Prior to Hyperpilot, Tim was an early employee at Mesosphere and CloudFoundry.
Tim is also active in the open-source space as an Apache member.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Nathan on LinkedIn: https://www.linkedin.com/in/nathanbenaich/
Connect with Tim on LinkedIn: https://www.linkedin.com/in/timchen
Timestamps:
0:00 - Nathan Benaich & Timothy Chen
1:36 - Tim's background
4:07 - Nathan's background
8:08 - To Nathan: What's your take on the lay of the land in the MLOps fear or space?
10:20 - To Tim: Can you give us your rundown on what you've been seeing? The greater landscape that you look at.
14:35 - To Tim: What companies right now really excite you? What are some that are doing something that has a future?
19:36 - To Nathan: What kind of companies you're looking at right now that you're doing interesting things?
22:37 - The MLOps tools mature as the companies mature.
23:45 - There's no tool that looks exactly the same from MLOps prospective
25:44 - Sometimes MLOps tools is not a choice by data scientists at all.
28:10 - What MLOps needs that are not being addressed by the market right now?
35:00 - What is the annotation stack?
37:28 - How do you think about in the context of federated learning?
41.24 - Will MLOps tools eventually become idiomatic? Would that be desirable?
47:55 - How do you switch from this open-source model to the money-making model?
52:30 - Should we focus only on the open-source only at first and think about monetization later? If so, are investors prepared to invest in no revenue companies?
**AI and ethical dilemmas**
Artificial Intelligence is seen by many as a vehicle for great transformation, but for others, it still remains a mystery, and many questions remain unanswered: will AI systems rule us one day? Can we trust AI to rule our criminal systems? Maybe create political campaigns and dominate political advertisements? Or maybe something less harmful, do our laundry? Some of these questions may sound absurd, but they are for sure making people shift from thinking purely about functional AI capabilities but also to look further to the ethics behind creating such powerful solutions.
For this episode we count with Charles Radclyffe as a guest, the data philosopher, to cover some of these dilemmas. You can reach out to Charles through LinkedIn or at ethicsgrade.io
Useful links:
- MLOps.Community slack
- TEDx talk - Surviving the Robot Revolution
- Digital Ethics whitepaper
MLOps community meetup #42! Last Wednesday, we talked to Mark Craddock, Co-Founder & CTO, Global Certification and Training Ltd (GCATI), about UN Global Platform.
// Abstract:
Building a global big data platform for the UN. Streaming 600,000,000+ records / day into the platform. The strategy developed using Wardley Maps and the Platform Design Toolkit.
// Bio:
Mark contributed to the Cloud First policy for the UK Public sector and was one of the founding architects for the UK Governments G-Cloud programme. Mark developed the initial CloudStore which enabled the UK Public Sector to procure cloud services from over 2,500 suppliers. The UK Public Sector has now purchased over £6.3Bn of cloud services, with £3.6Bn from Small to Medium Enterprises in the UK.
Mark lead the development of the United Nations Global Platform. A multi-cloud platform for capacity building within the national statistics offices in the use of Big Data and its integration with administrative sources, geospatial information, traditional survey and census data.
Mark is now building a non-profit training and certification organization.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Mark on LinkedIn: https://www.linkedin.com/in/markcraddock/
Timestamps:
[0:00] - Intro to Mark Craddock
[03:35] - Mark's background
[05:05] - UN Global Platform
[05:18] - Vision: A global collaboration to harness the power of data for better lives
[05:37] - UN GWG (Big) Data Membership
[05:49] - Sustainable Development Goals
[06:21] - Using the platform
[06:30] - Approach
[06:44] - Principles
[07:29] - How big was the team who put this together?
[08:09] - Leave no one behind. Endeavour to reach the furthest behind first.
[08:24] - Platform Business Model
[10:06] - Six distinct aspects of a platform and its ecosystem
[10:46] - The platform is the only business model able to orchestrate the wide range of products and services in an ecosystem
[11:09] - Through the means of a platform organization, ecosystems are capable of providing an improbable combination of attributes
[11:55] - Platforms and business models are also one of the best organizational structures for enabling rapid evolution
[13:22] - Technology Strategy
[13:23] - Wardley Maps
[14:50] - Is this were Machine Learning tools would fit in?
[20:35] - Are you looking how fast these are moving across to the right? How can you gauge that?
[26:57] - Is the value fluid?
[28:43] - How did you factor in the different personas?
[30:34] - How do you enable loosely coupled teams?
[35:44] - Data also moves from left to right
[42:00] - Technology Strategy Handbook
[42:20] - Achievements - July '19
[42:31] - Global Billing Intelligence
[43:15] - Privacy-Preserving Techniques Handbook
[43:26] - Cryptographic Techniques
[44:12] - Global Big Datasets
[44:55] - Big Data
[47:41] - Automatic Identification System (AIS)
[48:14] - Automatic Dependent Surveillance (ADS-B)
[48:41] - Satellite Imagery
[49:11] - Services in the platform
[49:16] - Location Analytics Service
[50:06] - Stack Sample
[50:37] - Data Sources
[51:50] - NiFi Dataflow
[52:20] - Is this how you enabled reproducibility?
[53:47] - Location Analytics Service
[55:31] - Shanghai - Flights
[55:45] - Shanghai - Cargo Ships
[56:00] - UN Global Platform
What are regulations saying about data privacy?
We are already aware of the importance of using Machine Learning to improve businesses, nevertheless to feed Machine Learning, data is a must, and in many cases, this data might even be considered sensitive information. So, does this mean that with new privacy regulations, access to data will be more and more difficult? ML and Data Science have their days counted? Or Will Machine beat privacy?
To answer all these questions I’ve invited Cat Coode, an expert on Data Privacy regulations, to join me in this episode, and help us sort out these questions!
Don’t forget to subscribe to the Mlops.community slack and if you’re looking for privacy-preserving solutions, show us some love and give a star to the Synthetic data open-source repo (https://github.com/ydataai/ydata-synthetic)
Useful links:
In this episode, we talked to Elizabeth Chabot, Consultant at Deloitte, about When You Say Data Scientist Do You Mean Data Engineer? Lessons Learned From StartUp Life.
// Key takeaways:
If you have a data product that you want to function in production, you need MLOps Education needs to happen about the data product life cycle, noting that ML is just part of the equation Titles need to be defined to help outside users understand the differences in roles
// Abstract:
ML and AI may sound sexy to investors, but if you work in the field you've probably spent late nights reviewing outputs manually, poured over logs and ran root cause analyses until your eyes hurt. If you've created data products at a company where analytics and data science held no meaning before your arrival, you've probably spent many-a-late-night explaining the basics of data collection, why ETL cannot be half-baked and that when you create a supervised model it needs to be supervised. Companies hoping to create a data product can have a data scientist show them how ML/AI can further their product, help them scale, or create better recommendations than their competitors. What companies are not always aware of is once the algorithm is created the data scientist is usually handicapped until more data-hires are made to build the necessary pipelines and frontend to put the algorithm in production. With the number of unique data-titles growing each year, how should the first data-evangelist-wrangler-wizard navigate title assignment?
// Bio: Elizabeth is a researcher turned data nerd. With a background in social and clinical sciences, Elizabeth is focused on developing data solutions that focus on creating value adds while allowing the user to make more intelligent decisions.
----------- Connect With Us ✌️-------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
MLOps community meetup #41! Last Wednesday was an exciting episode that some attendees couldn't help to ask when is the next season of their favorite series! The conversation was around Metaflow: Supercharging Data Scientist Productivity with none other than Netflix’s very own Ravi Kiran Chirravuri.
// Abstract:
Netflix's unique culture affords its data scientists an extraordinary amount of freedom. They are expected to build, deploy, and operate large machine learning workflows autonomously without the need to be significantly experienced with systems or data engineering. Metaflow, our ML framework (now open-source at metaflow.org), provides them with delightful abstractions to manage their project's lifecycle end-to-end, leveraging the strengths of the cloud: elastic compute and high-throughput storage. In this talk, we preface with our experience working alongside data scientists, present our human-centric design principles when building Machine Learning Infrastructure, and showcase how you can adopt these yourself with ease with open-source Metaflow.
// Bio:
Ravi is an individual contributor to the Machine Learning Infrastructure (MLI) team at Netflix. With almost a decade of industry experience, he has been building large-scale systems focusing on performance, simplified user journeys, and intuitive APIs in MLI and previously Search Indexing and Tensorflow at Google.
----------- Connect With Us ✌️-------------
Join our Slack community:
https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Ravi on LinkedIn: https://www.linkedin.com/in/seeravikiran/
Timestamps:
[00:00] - Introduction to Ravi Kiran Chirravuri
[02:21] - Ravi's background
[05:19] - Metaflow: Supercharging Data Scientist Productivity
[05:31] - Why do we have to build Metaflow?
[06:14] - Infographic of a very simplified view of a machine learning workflow
[07:01] - "An idea is typically meaningless without execution."
[07:38] - Scheduling
[08:14] - Life is great!
[08:24] - Life happens and things are crashing and burning!
[09:04] - What is Metaflow?
[12:01] - How much data scientist cares
[12:25] - How infrastructure is needed
[13:03] - What Metaflow does
[13:44] - How can you go about using Metaflow for your data science needs?
[14:20] - People love DAG's
[16:00] - Baseline
[16:16] - Architecture
[17:28] - Syntax
[19:00] - Vertical Scalability
[21:10] - Horizontal Scalability
[22:59] - Failures are a feature
[23:57] - State Transfer and Persistence
[27:05] - Dependencies
[30:57] - Model Ops: Versioning
[33:19] - Monitoring in Notebooks
[35:16] - Decouple Orchestration
[36:48] - AWS Step Functions
[37:16] - Export to AWS Step Functions
[38:10] - From Prototype to Production and Back
[42:07] - What are the prerequisites to use Metaflow?
[43:32] - Where does Metaflow store everything?
[45:10] - Are there any tutorials available?
[45:22] - Have the tutorials been updated?
[47:27] - How do you deploy Metaflow?
[49:02] - Do you see Metaflow becoming a tool to develop and support auto ML.
[50:34] - What were some of the biggest learnings that you saw people doing that they're not doing on Netflix?
[52:19] - Does Metaflow exist to help data scientists to orchestrate everything?
[54:30] - What do you version?
Coffee Sessions #18 with Luigi Patruno of ML in Production, a Centralized Repository of Best Practices
Summary
Luigi Patruno and ML in production
MLOps workflow: Knowledge sharing and best practices
Objective: learn!
Links:
ML in production: https://mlinproduction.com/
Why you start MLinProduction: https://mlinproduction.com/why-i-started-mlinproduction/
Luigi Patruno: a man whose goal is to help data scientists, ML engineers, and AI product managers, build and operate machine learning systems in production.
Luigi shares with us why he started ML in Production - A lot irrelevant content; a lot of clickbait with low standards of quality.
He had an Entrepreneurial itch and The solution was to start a weekly newsletter. From there he started creating Blog posts and now teamed up with Sam Charrington of TWIML to create courses on SagMaker ML.
Applied ML
Best practices
Reading google and microsoft papers
Analyzing the tools that are out there ie sagemaker and how to the see the world?
Aimed at making you more effective and efficient at your job
Community questions
Taking some time to answer some community questions!
Who do you learn from? Favorite resources?
Self-taught, papers, talks
Construct the systems
Uber michelangelo
----------------- 📝 Rought notes 📝 ----------------
Any companies that stand out to you in terms of MLOps excellence?
Google, Amazon, Stichfix: they've had to solve hard problems
Serving ads
Personalization at scale
Vertical problems: within their vertices
Motivated by real challenges
DropBox
Great articles
A great machine learning company
Tools
Sagemaker
Has a course on sagemaker
Nice lessons baked into the system
Dos and don’t of MLOps
DO LOG!
Monitor
Automate - manual analysis leads to problems
Do it manually first til you feel confident that you can automate it
Tag, version
Store your training, val, and test sets!
What is his process of identifying use cases that are suitable for machine learning as a solution? How do they proceed methodically?
Start with business goal
Potential number of users that the solution can benefit
The ability to build a predictive model
Performance x impact = score
Rank problems by this
How developed are the datasets?
What part of the ML in Production process do people underestimate the most? What are the low hanging fruits that many people don’t take advantage of?
Generate actual value without needing to build the most complex model possible
In industry, performance is only one part of the equation
How has he seen ML in production evolve over the last few years and where does he think it's headed next?
More and more tools!
Industry-specific tool taking advantage of ML
Problem is you must have industry knowledge
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
This is the first episode of a podcast series on Machine Learning and Data privacy. Machine Learning is the key to the new revolution in many industries. Nevertheless, ML does not exist without data and a lot of it, which in many cases results in the use of sensitive information. With new privacy regulations, access to data is today harder and much more difficult but, does that mean that ML and Data Science has its days counted? Will the Machines beat privacy?
Don’t forget to subscribe to the mlops.community slack (https://go.mlops.community/slack) and to give a star to the Synthetic data open-source repo (https://github.com/ydataai/ydata-synt...)
Useful links:
Medium post with the podcast transcription - https://medium.com/@fabiana_clemente/...
In case you’re curious about GDPR fines - enforcementtracker.com
The Netflix Prize - https://www.nytimes.com/2010/03/13/technology/13netflix.html
Tensorflow privacy - https://github.com/tensorflow/privacy
MLOps level 2: CI/CD pipeline automation
For a rapid and reliable update of the pipelines in production, you need a robust automated CI/CD system. This automated CI/CD system lets your data scientists rapidly explore new ideas around feature engineering, model architecture, and hyperparameters. They can implement these ideas and automatically build, test, and deploy the new pipeline components to the target environment.
Figure 4. CI/CD and automated ML pipeline.
This MLOps setup includes the following components:
Source control
Test and build services
Deployment services
Model registry
Feature store
ML metadata store
ML pipeline orchestrator
Characteristics of stages discussion.
Figure 5. Stages of the CI/CD automated ML pipeline.
The pipeline consists of the following stages:
Development and experimentation: You iteratively try out new ML algorithms and new modelling where the experiment steps are orchestrated. The output of this stage is the source code of the ML pipeline steps that are then pushed to a source repository.
Pipeline continuous integration: You build source code and run various tests. The outputs of this stage are pipeline components (packages, executables, and artefacts) to be deployed in a later stage.
Pipeline continuous delivery: You deploy the artefacts produced by the CI stage to the target environment. The output of this stage is a deployed pipeline with the new implementation of the model.
Automated triggering: The pipeline is automatically executed in production based on a schedule or in response to a trigger. The output of this stage is a trained model that is pushed to the model registry.
Model continuous delivery: You serve the trained model as a prediction service for the predictions. The output of this stage is a deployed model prediction service.
Monitoring: You collect statistics on the model performance based on live data. The output of this stage is a trigger to execute the pipeline or to execute a new experiment cycle. The data analysis step is still a manual process for data scientists before the pipeline starts a new iteration of the experiment. The model analysis step is also a manual process.
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
MLOps community meetup #40! Last Wednesday, we talked to Theofilos Papapanagiotou, Data Science Architect at Prosus, about Hands-on Serving Models Using KFserving.
// Abstract:
We looked to some popular model formats like the SavedModel of Tensorflow, the Model Archiver of PyTorch, pickle&ONNX, to understand how the weights of the NN are saved there, the graph, and the signature concepts.
We discussed the relevant resources of the deployment stack of Istio (the Ingress gateway, the sidecar and the virtual service) and Knative (the service and revisions), as well as Kubeflow and KFServing. Then we got into the design details of KFServing, its custom resources, the controller and webhooks, the logging, and configuration.
We spent a large part in the monitoring stack, the metrics of the servable (memory footprint, latency, number of requests), as well as the model metrics like the graph, init/restore latencies, the optimizations, and the runtime metrics which end up to Prometheus. We looked at the inference payload and prediction logging to observe drifts and trigger the retraining of the pipeline.
Finally, a few words about the awesome community and the roadmap of the project on multi-model serving and inference routing graph.
// Bio:
Theo is a recovering Unix Engineer with 20 years of work experience in Telcos, on internet services, video delivery, and cybersecurity. He is also a university student for life; BSc in CS 1999, MSc in Data Coms 2008, and MSc in AI 2017.
Nowadays he calls himself an ML Engineer, as he expresses through this role his passion for System Engineering and Machine Learning.
His analytical thinking is driven by curiosity and hacker spirit. He has skills that span a variety of different areas: Statistics, Programming, Databases, Distributed Systems, and Visualization.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Theofilos on LinkedIn: https://linkedin.com/in/theofpa
MLOps community meetup #39! Last week we talked to Ivan Nardini, Customer Engineer at SAS, about Operationalize Open Source Models with SAS Open Model Manager.
// Abstract:
Analytics are Open.
According to their nature, Open Source technologies allows an agile development of the models, but it results difficult to put them in production. The goal of SAS is supporting customers in operationalize analytics In this meetup,
I present SAS Open Model Manager, a containerized Modelops tool that accelerates deployment processes and, once in production, allows monitoring your models (SAS and Open Source).
// Bio:
As a member of Pre-Sales CI & Analytics Support Team, I'm specialized in ModelOps and Decisioning. I've been involved in operationalizing analytics using different Open Source technologies in a variety of industries. My focus is on providing solutions to deploy, monitor and govern models in production and optimize business decisions processes. To reach this goal, I work with software technologies (SAS Viya platform, Container, CI/CD tools) and Cloud (AWS).
//Other Links you can check Ivan on:
https://medium.com/@ivannardini
----------- Connect With Us ✌️-------------
Join our Slack community:
https://go.mlops.community/slack
Follow us on Twitter:
@mlopscommunity
Sign up for the next meetup:
https://go.mlops.community/register
Connect with Demetrios on LinkedIn:
https://www.linkedin.com/in/dpbrinkm/
Connect with Ivan on LinkedIn:
https://www.linkedin.com/in/ivan-nardiniDescription
Timestamps: 0:00 - Intro to Ivan Nardini
3:41 - Operationalize Open Source Models with SAS Open Model Manager slide
4:21 - Agenda
5:01 - What is ModelOps and what is the difference between MLOps and ModelOps?
6:19 - "Do I look like an expert?" Ivan's Background
7:12 - Why ModelOps?
7:20 - Operationalizing Analytics
8:12 - Operationalizing Analytics: SAS
9:08 - Operationalizing Analytics: Customer
11:36 - What's a model for you?
12:07 - Hidden Complexity in ML Systems
12:52 - Hidden Complexity in ML Systems: Business Prospective
14:12 - Hidden Complexity in ML Systems: IT Prospective
17:12 - One of the hardest things is Security?
17:52 - Hidden Complexity in ML Systems: Analytics Prospective
19:20 - Why ModelOps?
20:09 - ModelOps technologies Map
22:29 - Customers ModelOps Maturity over Technology Propensity. MLOps Maturity vs. Technology Propensity
26:23 - Show us your Analytical Models
26:56 - SAS can support you to ship them in production providing Governance and Decisioning.
27:28 - When you talk to people, is there something that you feel like there is a unified model, but focusing on the wrong thing?
29:14 - Have you seen Reproducibility and Governance?
30:47 - Advertising Time
30:55 - Operationalize Open Source Models with SAS Open Model Manager
31:02 - ModelOps with SAS
32:06 - SAS Open Model Manager
33:18 - Demo
33:27 - SAS Model Ops Architecture - Classification Model
35:02 - Model Demo: Credit Scoring Business Application
50:20 - Take Homes
50:24 - Operationalize Analytics
50:32 - Model Lifecycle Effort Side
51:20 - Business Value Side
51:47 - Typical Analytics Operationalization Graph
52:18 - Analytics Operationalization with ModelOps Graph
53:18 - Is this for everybody?
//Bio
Satish built compilers, profilers, IDEs, and other dev tools for over a decade. At Microsoft Research, he saw his colleagues solving hard program analysis problems using Machine Learning. That is when he got curious and started learning. His approach to ML is influenced by his software engineering background of building things for production. He has a keen interest in doing ML in production, which is a lot more than training and tuning the models. The first step is to understand the product and business context, then building an efficient pipeline, then training models, and finally monitoring its efficacy and impact on the business. He considers ML as another tool in the software engineering toolbox, albeit a very powerful one. He is a co-founder of Slang Labs, a Voice Assistant as a Service platform for building in-app voice assistants.
//Talk Takeaways ML-driven product features will grow manifold. Organizations take an evolutionary approach to absorb tech innovations. ML will be no exception. How Organizations adopted cloud can offer useful lessons.
ML/DS folks who invest in an understanding business context and tech environment of the org will make a bigger impact.
Organizations that invest in data infrastructure will be more successful in extracting value from machine learning.
//Other links you can check Satish on
An Engineer’s trek into Machine Learning:
https://scgupta.link/ml-intro-for-developers
Architecture for High-Throughput Low-Latency Big Data Pipeline on Cloud:
https://scgupta.link/big-data-pipeline-architecture
Data pipeline article:
https://scgupta.link/big-data-pipeline-architecture or
https://towardsdatascience.com/scalable-efficient-big-data-analytics-machine-learning-pipeline-architecture-on-cloud-4d59efc092b5
Tips for software engineers based on my experience of getting into ML:
https://scgupta.link/ml-intro-for-developers or https://towardsdatascience.com/software-engineers-trek-into-machine-learning-46b45895d9e0
Linkedin:
https://www.linkedin.com/in/scgupta
Twitter:
https://twitter.com/scgupta
Personal Website:
http://scgupta.me
Company Website:
https://slanglabs.in
Voice Assistants info:
https://www.slanglabs.in/voice-assistants
Timestamps:
0:00 - Intro to Satish Chandra Gupta
1:05 - Background of Satish on Machine Learning
3:29 - Satish's background on what he's doing now
5:34 - Why were you interested in the challenges of the workload?
9:53 - As you're looking at the data pipeline, do you see much overlap there?
15:38 - Relationships between engineering pipeline characteristics and how they relate to data.
20:24 - Tips for saving when you're building these pipeline.
24:44 - First point of engagement: Collection
31:26 - Possibilities of Data Architecture
38:03 - Why is it beneficial to save money?
44:22 - Learnings of Satish with his current project, Voice Assistant as a service.
James Sutton is an ML Engineer focused on helping enterprise bridge the gap between what they have now, and where they need to be to enable production scale ML deployments.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with James on LinkedIn: https://www.linkedin.com/in/jamessutton2/
Timestamps:
0:00 - Intro to Speaker
2:20 - Scope of the coffee session
3:10 - Background of James Sutton
8:28 - One-shots Classifier Algorithm
12:46 - Why is it a challenge from the engineering perspective with deployment?
19:20 - How to overcome bottlenecks?
30:07 - Vision of your landscape?
34:45 - Maturity playout
38:48 - Maturity perspective of ML
41:49 - Risk of overgeneralizing system designs patterns
46:10 - Reliability, Speed, Cost
46:46 - Consistency, Availability, Partition Tolerance (CAP Theorem)
47:36 - How do you go about discussing these tradeoffs with your clients?
51: 23 - How would you deal with the PII?
58:50 - Collaborative process with clients
1:00:55 - Wrap up
Parallel Computing with Dask and Coiled
Python makes data science and machine learning accessible to millions of people around the world. However, historically Python hasn't handled parallel computing well, which leads to issues as researchers try to tackle problems on increasingly large datasets. Dask is an open source Python library that enables the existing Python data science stack (Numpy, Pandas, Scikit-Learn, Jupyter, ...) with parallel and distributed computing. Today Dask has been broadly adopted by most major Python libraries, and is maintained by a robust open source community across the world.
This talk discusses parallel computing generally, Dask's approach to parallelizing an existing ecosystem of software, and some of the challenges we've seen in deploying distributed systems.
Finally, we also addressed the challenges of robustly deploying distributed systems, which ends up being one of the main accessibility challenges for users today. We hope that by the end of the meetup attendees will better understand parallel computing, have built intuition around how Dask works, and have the opportunity to play with their own Dask cluster on the cloud.
Matthew is an open source software developer in the numeric Python ecosystem. He maintains several PyData libraries, but today focuses mostly on Dask a library for scalable computing. Matthew worked for Anaconda Inc for several years, then built out the Dask team at NVIDIA for RAPIDS, and most recently founded Coiled Computing to improve Python's scalability with Dask for large organizations.
Matthew has given talks at a variety of technical, academic, and industry conferences. A list of talks and keynotes is available at (https://matthewrocklin.com/talks).
Matthew holds a bachelor’s degree from UC Berkeley in physics and mathematics, and a PhD in computer science from the University of Chicago.
Check out our posts here to get more context around where we're coming from:
https://medium.com/coiled-hq/coiled-dask-for-everyone-everywhere-376f5de0eff4
https://medium.com/coiled-hq/the-unbearable-challenges-of-data-science-at-scale-83d294fa67f8
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Matthew on LinkedIn: https://www.linkedin.com/in/matthew-rocklin-461b4323/
This time we talked about one of the most vibrant questions for any MLOps practitioner: how to choose the right tools for your ML team, given the huge amount of open-source and proprietary MLOps tools available on the market today.
We discussed several criteria to rely on when choosing a tool, including:
- The requirements of the particular team use-cases
- The scaling capacity of the tool
- The cost of migration from a chosen tool
- The cost of teaching the team to use this tool
- The company or the community behind the tool
Apart from that, we talked about particular use-cases and discussed the trade-offs between waiting for a new release of your tool to get the missing piece of functionality, switching to another tool, and building an in-house solution.
We also touched the topic of organising MLOps teams and practices across large companies with a lot of ML teams.
// Bio:
Jose Navarro
Jose Navarro is a Machine Learning Infrastructure Engineer making everyday cooking fun at Cookpad, where its recipe platform has more than 40 million monthly users. He holds a MSc in Machine Learning and High-Performance Computing from the University of Bristol. He is interested in Cloud Native technologies, serverless, and event-driven architecture.
Mariya Davidova
Mariya came to MLOps from a software development background. She started her career as a Java developer in JetBrains in 2011, then gradually moved to developer advocacy for JS-based APIs. In 2019, she joined Neu.ro as a platform developer advocate and then moved to the product management position.
Mariya has been obsessed with AI and ML for many years: she finished a bunch of courses, read a lot of books, and even wrote a couple of fiction stories about AI. She believes that proper tooling and decent development and operations practices are essential success component for ML projects, as well as they are for traditional SD.
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Jose on LinkedIn: https://www.linkedin.com/in/jose-navarro-2a57b612/
Connect with Maria on LinedIn: https://www.linkedin.com/in/mariya-davydova/
Dask
What is it?
Parallelism for analytics
What is parallelism?
Doing a lot at once by splitting tasks into smaller subtasks which can be processed in parallel (at the same time)
Distributed work across multiple machines and then combining the results
Helpful for CPU bound - doing a bunch of calculations on the CPU. The rate at which process progresses is limited by the speed of the CPU
Concurrency?
Similar but a but things don’t have to happen at the same time, they can happen asynchronously. They can overlap.
Shared state
Helpful to I/O bound - networking, reading from disk, etc. The rate at which a process progresses is limited by the speed of the I/O subsystem.
Multi-core vs distributed
Multi-core is a single processor with 2 or more cores that can cooperate through threads - multithreading
Distributed is across multiple nodes communicating via HTTP or RPC Why is this hard?
Python has it challenges due to GIL, other languages don't have this problem
Shared state can lead to potential race conditions, deadlocks, etc
Coordination work across the machines
For analytics?
Calculating some statistics on a large dataset can be tricky if it can’t fit in memory
// Show Notes
Coiled Cloud: https://cloud.coiled.io/
Coiled Launch Announcement: https://medium.com/coiled-hq/coiled-dask-for-everyone-everywhere-376f5de0eff4
OSS article: https://www.forbes.com/sites/glennsolomon/2020/09/15/monetizing-open-source-business-models-that-generate-billions/#2862e47234fd
Amish barn raising: https://www.youtube.com/watch?v=y1CPO4R8o5M
MessagePassingInterface: https://en.wikipedia.org/wiki/Message_Passing_Interface
----------- Connect With Us ✌️-------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Matthew on LinkedIn: https://www.linkedin.com/in/matthew-rocklin-461b4323/
Timestamps:
0:00 - Intro to Matthew Rocklin and Hugo Bowne-Anderson
0:37 - Matthew Rocklin's Background
1:17 - Hugo Brown-Anderson's Background
3:47 - Where did that inspiration come from?
10:04 - Is there a close relationship between Best Practices and Tooling or are these two separate things?
11:27 - Why is Data Literacy important with Coiled?
14:46 - How do you think about the balance between enabling Data Science to have a lot of powerful compute?
17:05 - Machine Learning as a space for tracking best practices experimentation
19:32 - What makes Data Science so difficult?
24:07 - How can a for-profit company compliment Open Source Software (OSS)
29:40 - Amazon becoming a competitor with your own open-source technology (?)
32:50 - How do you encourage more people to contribute and ensure quality?
34:58 - Do you see Coiled operating within the DASK ecosystem?
37:30 - What is DASK?
39:19 - What should people know about parallelism?
41:28 - Why is it so hard to put things back together?
41:34 - Why does Python need a whole new tool to enable that? Or maybe some other tools as well?
44:44 - Dynamic Tasks Scheduling as being useful to Data Scientists
47:15 - Why is reliability in particular important in Data Science?
52:27 - What's in store for DASK?
Why was Flyte built at Lyft?
What sorts of requirements does a ML infrastructure team have at lyft?
What problems does it solve / use cases?
Where does it fit in in the ML and Data ecosystem?
What is the vision?
Who should consider using it?
Learnings as the engineering team tried to bootstrap an open-source community.
Ketan Umare is a senior staff software engineer at Lyft responsible for technical direction of the Machine Learning Platform and is a founder of the Flyte project. Before Flyte he worked on ETA, routing and mapping infrastructure at Lyft. He is also the founder of Flink Kubernetes operator and contributor to Spark on kubernetes. Prior to Lyft he was a founding member of Oracle Baremetal Cloud and lead teams building Elastic Block Storage. Prior to that, he started and lead multiple teams in Mapping and Transportation optimization infrastructure at Amazon. He received his Masters in Computer Science from Georgia Tech specializing in High-performance computing and his Bachelors in Engineering in Computer Science from VJTI Mumbai.
Besides work, he enjoys spending time with his daughter and wife. He loves the Pacific Northwest outdoors and will try anything new.
Lyft
Pricing, Locations, Estimated Time of Arrivals (ETA), Mapping, Self-Driving (L5), etc.
What sort of scale, storage, network bandwidth are we looking at?
Tens of thousands of workflows, hundreds of thousands of executions, millions of tasks, and tens of millions of containers!
Flyte: more than 900k workflow executed a month and more than 30+ million container executions per month
Typical flow of information?
What are the user stories you’re typically dealing with at lyft?
How do you set it up?
On-prem, cloud, etc.
Helm installable?
Why Golang?
What problems does it solve?
Complex data dependencies? Why
Orchestrated compute on demand
Reuse and sharing
Key features
Multi-tenant, hosted, serverless
Parametrized, data lineage, caching
Additionally, if the run invokes a task that has already been computed before, regardless of who executed it, Flyte will smartly use the cached output, saving you both time and money.
Versioning, sharing
Modular, loosely coupled
Seems like you guys recognize that the best task for the job might be hosted elsewhere, so it was important to integrate other solutions into flyte.
Flyte extensions
Backend plugins - is it true you can create and manage k8s resources like CRDs for things like spark, sagemaker, bigquery?
Drop a Star
https://flyte.org
Flyte community
----------- Connect With Us ✌️-------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Ketan on LinkedIn: https://www.linkedin.com/in/ketanumare/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Round 3 analyzing the Google paper "Continuous Delivery and Automation Pipelines in ML"
// Show Notes
Data Science Steps for ML
Data extraction: You select and integrate the relevant data from various data sources for the ML task.
Data analysis: You perform exploratory data analysis (EDA) to understand the available data for building the ML model. This process leads to the following:
Understanding the data schema and characteristics that are expected by the model.
Identifying the data preparation and feature engineering that are needed for the model.
Data preparation: The data is prepared for the ML task. This preparation involves data cleaning, where you split the data into training, validation, and test sets. You also apply data transformations and feature engineering to the model that solves the target task. The output of this steps are the data splits in the prepared format.
Model training: The data scientist implements different algorithms with the prepared data to train various ML models. In addition, you subject the implemented algorithms to hyperparameter tuning to get the best performing ML model. The output of this step is a trained model.
Model evaluation: The model is evaluated on a holdout test set to evaluate the model quality. The output of this step is a set of metrics to assess the quality of the model.
Model validation: The model is confirmed to be adequate for deployment—that its predictive performance is better than a certain baseline.
Model serving: The validated model is deployed to a target environment to serve predictions. This deployment can be one of the following:
Microservices with a REST API to serve online predictions.
An embedded model to an edge or mobile device.
Part of a batch prediction system.
Model monitoring: The model predictive performance is monitored to potentially invoke a new iteration in the ML process.
The level of automation of these steps defines the maturity of the ML process, which reflects the velocity of training new models given new data or training new models given new implementations. The following sections describe three levels of MLOps, starting from the most common level, which involves no automation, up to automating both ML and CI/CD pipelines.
In the rest of the conversation, we talk about maturity levels 0 and 1. Next session we will talk about Level 2.
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
MLOps community meetup #36! This week we talk to David Hershey Solutions Engineer at Determined AI, about Moving Deep Learning from Research to Production with Determined and Kubeflow.
// Key takeaways:
What components are needed to do inference in ML
How to structure models for ML inference
How a model registry helps organize your models for easy consumption
How you can set up reusable and easy-to-upgrade inference pipelines
// Abstract:
Translating the research that goes into creating a great deep learning model into a production application is a mess without the right tools. ML models have a lot of moving pieces, and on top of that models are constantly evolving as new data arrives or the model is tweaked. In this talk, we'll show how you can find order in that chaos by using the Determined Model Registry along with Kubeflow Pipelines.
// Bio:
David Hershey is a solutions engineer for Determined AI. David has a passion for machine learning infrastructure, in particular systems that enable data scientists to spend more time innovating and changing the world with ML. Previously, David worked at Ford Motor Company as an ML Engineer where he led the development of Ford's ML platform. He received his MS in Computer Science from Stanford University, where he focused on Artificial Intelligence and Machine Learning.
// Relevant Links
www.determined.ai
https://github.com/determined-ai/determined
https://determined.ai/blog/production-training-pipelines-with-determined-and-kubeflow/
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
https://www.linkedin.com/in/david-hershey-458ab081/
Timestamps:
0:00 - Intros
4:15 - The structure of the chat
5:20 - What is DeterminedAI?
7:20 - How is DeterminedAI different than other more standard artifact storage solutions?
9:25 - Where are the boundaries between what your tool determined AI does really well, and where it works smoothly with other things around it?
11:48 - Is Kubeflow dying?
13:54 - How do you see DeterminedAI and Kubeflow becoming more solidified?
15:55 - How does DeterminedAI interact with Kubeflow at the moment?
18:01 - What type of models they are, is the Kubeflow metadata?
19:18 - What a model registry is and why it's so important to have that?
23:16 - Can you give us the quick demo real fast?
30:52 - Which orchestration tool to use?
32:04 - When using Kubeflow are determined how can you deploy the model through CD tools like Jenkins?
33:40 - How is determined connected to Kubeflow?
36:09 - What components you feel are needed to do inference in machine learning? And how can we structure different models for that machine learning inference?
40:04 - Are they the same one when we talk about ML researchers?
42:14 - How can we better be ready for when we do want to get into the production?
44:59 - In this pipeline, Where do you normally see people getting stopped?
47:05 - What are things that you've been seen pop up that you're not necessarily thinking about in those first phases?
50:17 - What are the most underrated topic regarding deploying machine learning models in production?
52:44 - How do you see the adoption of tools such as Determined and Kubeflow by Data scientists?
54:40 - Can you explain the Determined open source components?
Second installation David and Demetrios reviewing the google paper about Continuous training and automated pipelines. They dive deep into machine learning monitoring and also what exactly continuous training actually entails. Some key highlights are:
Automatically retraining and serving the models:
When to do it?
Outlier detection
Drift detection
Outlier detection:
What is it?
How you deal with it
Drift detection
Individual features may start to drift. This could be a bug or it could be perfectly normal behavior that indicates that the world has changed requiring the model to be retrained.
Example changes:
shifts in people’s preferences
marketing campaigns
competitor moves
the weather
the news cycle
Locations
Time
Devices (clients)
If the world you're working with is changing over time, model deployment should be treated as a continuous process. What this tells me is that you should keep the data scientists and engineers working on the model instead of immediately moving to another project.
Deeper dive into concept drift
Feature/target distributions change
An overview of concept drift applications: “.. data analysis applications, data evolve over time and must be analyzed in near real time. Patterns and relations in such data often evolve over time, thus, models built for analyzing such data quickly become obsolete over time. In machine learning and data mining this phenomenon is referred to as concept drift.”
https://www.win.tue.nl/~mpechen/publications/pubs/CD_applications15.pdf
https://www-ai.cs.tu-dortmund.de/LEHRE/FACHPROJEKT/SS12/paper/concept-drift/tsymbal2004.pdf
Types of concept drift:
Sudden
Gradual
Google in some way is trying to address this concern - the world is changing and you want your ML system to change as well so it can avoid decreased performance but also improve over time and adapt to its environment. This sort of robustness is necessary for certain domains.
Continuous delivery and automation of pipelines (data, training, prediction service) was built with this in mind. Minimizing the commit-to-deploy interval and maximize the velocity software delivery and its components: maintainability, extensibility, and testability
Then the pipeline is ready, you can now run it. So you can do this continuously. After the pipeline is deployed to the production environment, it will be executed automatically and repetitively to produce a trained model that is stored in a central model registry.
This pipeline should be able to be run on a schedule or based on triggers: certain events that you have configured to your business domain - new data or drop in performance from the prod model.
The link between the model artifact and the pipeline is never severed. What pipeline trained them? What data was extracted, validated and how was it prepared? What was the training configuration and how was it evaluated? Etc. metrics are key here! Lineage tracking!!!
Keeping a close tie between the dev/experiment pipeline and the continuous production pipeline helps avoid inconsistencies between model artifacts produced by the pipeline and models beings served - hard to debug
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
MLOps Meetup #34! This week we talk to Kai Waehner about the beast that is apache kafka and how many different ways you can use it!
// Key takeaways:
-Kafka is much more than just messaging
-Kafka is the de facto standard for processing huge volumes of data at scale in real-time
-Kafka and Machine Learning are complementary for various use cases (including data integration, data processing, model training, model scoring, and monitoring)
// Abstract:
The combination of Apache Kafka, tiered storage, and machine learning frameworks such as TensorFlow enables you to build one scalable, reliable, but also simple infrastructure for all machine learning tasks using the Apache Kafka ecosystem and Confluent Platform. This discussion features a predictive maintenance use case within a connected car infrastructure, but the discussed components and architecture are helpful in any industry.
// Bio:
Kai Waehner is a Technology Evangelist at Confluent. He works with customers across the globe and with internal teams like engineering and marketing. Kai’s main area of expertise lies within the fields of Big Data Analytics, Machine Learning, Hybrid Cloud Architectures, Event Stream Processing and Internet of Things. He is a regular speaker at international conferences such as Devoxx, ApacheCon and Kafka Summit, writes articles for professional journals, and shares his experiences with new technologies on his blog: www.kai-waehner.de.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Kai: [email protected] / @KaiWaehner / LinkedIn (https://www.linkedin.com/in/megachucky/)
________Show Notes_______
Blogpost tierd storage
https://www.confluent.io/blog/streaming-machine-learning-with-tiered-storage/
https://www.confluent.io/resources/kafka-summit-2020/apache-kafka-tiered-storage-and-tensorflow-for-streaming-machine-learning-without-a-data-lake/
Blogpost about using kafka as a database
https://www.kai-waehner.de/blog/2020/03/12/can-apache-kafka-replace-database-acid-storage-transactions-sql-nosql-data-lake/
Example repo on github
https://github.com/kaiwaehner/hivemq-mqtt-tensorflow-kafka-realtime-iot-machine-learning-training-inference
Model serving vs embedded kafka
https://www.confluent.io/blog/machine-learning-real-time-analytics-models-in-kafka-applications/
https://www.confluent.io/kafka-summit-san-francisco-2019/event-driven-model-serving-stream-processing-vs-rpc-with-kafka-and-tensorflow/
Istio blog post
https://www.kai-waehner.de/blog/2019/09/24/cloud-native-apache-kafka-kubernetes-envoy-istio-linkerd-service-mesh/
While machine learning is spreading like wildfire, very little attention has been paid to the ways that it can go wrong when moving from development to production. Even when models work perfectly, they can be attacked and/or degrade quickly if the data changes. Having a well understood MLOps process is necessary for ML security!
Using Kubeflow, we demonstrated how to the common ways machine learning workflows go wrong, and how to mitigate them using MLOps pipelines to provide reproducibility, validation, versioning/tracking, and safe/compliant deployment. We also talked about the direction for MLOps as an industry, and how we can use it to move faster, with less risk, than ever before.
David leads Open Source Machine Learning Strategy at Azure. This means he spends most of his time helping humans to convince machines to be smarter. He is only moderately successful at this. Previously, he led product management for Kubernetes on behalf of Google, launched Google Kubernetes Engine, and co-founded the Kubeflow project. He has also worked at Microsoft, Amazon and Chef and co-founded three startups. When not spending too much time in service of electrons, he can be found on a mountain (on skis), traveling the world (via restaurants) or participating in kid activities, of which there are a lot more than he remembers than when he was that age.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aronchick/
In this last episode, we covered how Google is thinking about MLOps and how automation plays a key part in their view of MLOps. We started to talk about CI, CD, and the role they play in a pipeline setup for CT. In the next episode, we'll pick up where we left off, starting our discussion of CT and some of the reasons you’d want to set up a pipeline with continuous training in the first place.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Yoav is the builder behind Say Less, an AI-powered email summarization tool that was recently featured on the front page of Hacker News and Product Hunt. In this talk, Yoav will walk us through the end-to-end process of building the tool, from the prototype phase to deploying the model as a realtime HTTP endpoint.
Yoav Zimmerman is the engineer / founder behind Model Zoo, a machine learning deployment platform focused on ease-of-use. He has previously worked at Determined AI on large-scale deep learning training infrastructure and at Google on knowledge base construction for features that powered Google Assistant and Search.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Yoav on LinkedIn: https://www.linkedin.com/in/yoav-zimmerman-05653252/
|| Links Referenced in the Show ||
General Info: https://medium.com/@paktek123
Load Balancer Series: https://medium.com/load-balancer-series
Upcoming Open Src: https://medium.com/upcoming-open-source
Some Libraries Neeran maintains: https://github.com/paktek123/elasticsearch-crystal
Some libraries Neeran used to maintain: https://github.com/microsoft/pgtester (and https://medium.com/yammer-engineering/testing-postgresql-scripts-with-rspec-and-pg-tester-c3c6c1679aec)
Some interesting projects Neeran has worked on (architected these): https://devblogs.microsoft.com/cse/2016/05/22/access-azure-blob-storage-from-your-apps-using-s3-api/, https://medium.com/yammer-engineering/logs-on-logs-on-logs-aggregation-at-yammer-2b7073f35606
Some of Benevolent Stuff: https://www.benevolent.com/engineering-blog/deploying-metallb-in-production, https://www.benevolent.com/engineering-blog/spark-on-kubernetes-for-nlp-at-scale (helped with the infra side)
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
We trained a Transformer neural net on ambient music to see if a machine can compose with the great masters. Ambient is a soft, flowing, ethereal genre of music that I’ve loved for decades. There are all kinds of ambient, from white noise, to tracks that mimic the murmur of soft summer rain in a sprawling forest, but Dan favors ambient that weaves together environmental sounds and dreamy, wavelike melodies into a single, lush tapestry.
Can machine learning ever hope to craft something so seemingly simple yet intricate? The answer is yes and it’s getting better and better with each passing year. It won’t be long before artists are co-composing with AI, using software that helps them weave their own masterpieces of sound.
In this talk, we looked at how we did it. Along the way we’ll listen to some more awesome samples that worked really well and some that didn’t work as well as we hoped. You can download the model to play around with yourself. Dam also shows you an end-to-end machine learning pipeline, with downloadable containers that you can string together with ease to train a masterful music-making machine learning model on your own.
Dan Jeffries is Chief Technology Evangelist at Pachyderm. He’s also an author, engineer, futurist, pro blogger and he’s given talks all over the world on AI and cryptographic platforms. He’s spent more than two decades in IT as a consultant and at open source pioneer Red Hat.
With more than 50K followers on Medium, his articles have held the number one writer's spot on Medium for Artificial Intelligence, Bitcoin, Cryptocurrency and Economics more than 25 times. His breakout AI tutorial series "Learning AI If You Suck at Math" along with his explosive pieces on cryptocurrency, "Why Everyone Missed the Most Important Invention of the Last 500 Years” and "Why Everyone Missed the Most Mind-Blowing Feature of Cryptocurrency,” are shared hundreds of times daily all over social media and been read by more than 5 million people worldwide.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Dan on LinkedIn: https://www.linkedin.com/in/danjeffries/
MLOps and DevOps have a large number of parallels. Many of the techniques, practices, and processes used for traditional software projects can be followed almost exactly in ML projects. However, the day-to-day of an ML project is usually significantly different from a traditional software project. So while the ideas and principles can still apply, it’s important to be aware of the core aims of DevOps when applying them.
Damian is a Cloud Advocate specializing in DevOps and MLOps. After spending a year in Toronto, Canada, he returned to Australia - the land of the dangerous creatures and beautiful beaches - in 2018. Formerly a dev at Octopus Deploy and a Microsoft MVP, he has a background in software development and consulting in a broad range of industries. In Australia, he co-organised the Brisbane .Net User Group, and launched the now annual DDD Brisbane conference. He regularly speaks at conferences, User Groups, and other events around the world. Most of the time you'll find him talking to software engineers, IT pros and managers to help them get the most out of their DevOps strategies.
|| Links Referenced in the Show ||
MLOps, or DevOps for Machine Learning: https://damianbrady.com.au/2019/10/28/mlops-or-devops-for-machine-learning/
Microsoft Azure Machine Learning: http://ml.azure.com/
MLOps Coffee Sessions #6 Continuous Integration for ML // Featuring Elle O'Brien: https://www.youtube.com/watch?v=L98VxJDHXMM
MLOps: Isn’t that just DevOps? Ryan Dawson speaks at MLOps Coffee Session: https://www.seldon.io/mlops-isnt-that-just-devops-ryan-dawson-speaks-at-mlops-coffee-session/
DVC - Data Version Control: https://dvc.org/
Pachyderm - Version-controlled data science: https://www.pachyderm.com/
Databricks - Unified Data Analytics: https://databricks.com/
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Damian on LinkedIn: https://www.linkedin.com/in/damianbrady/
The concept of machine learning products is a new one for the business world. There is a lack of clarity around key elements: Product Roadmaps and Planning, the Machine Learning Lifecycle, Project and Product Management, Release Management, and Maintenance.
In this talk, we covered a framework specific to Machine Learning products. We discussed the improvements businesses can expect to see from a repeatable process. We also covered the concept of monetization and integrating machine learning into the business model.
Vin is an applied data scientist and teaches companies to monetize machine learning. He is currently working on a ML based decision support product as well as my strategy consulting practice.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vin on LinkedIn: https://www.linkedin.com/in/vineetvashishta/
Timestamps:
[00:00] Intro to Vin Vashishta
[01:33] Vin's background
[05:04] Key problems when monetizing Machine Learning
[07:00] How can we fix the key problems in monetizing Machine Learning
[13:24] How can we go about creating that repeatable process?
[16:17] There are all these data scientists aren't going to school and getting all these diplomas for data wrangling. Right?
[17:12] How can you successfully envision that road mapping from the beginning of the process?
[24:19] How can a Data Scientist be more proactive instead of just getting paid?
[28:53] Have you figured out how to quickly estimate an order of magnitude when ROI questions arise?
[31:48] Have you seen a company that has machine learning as its core product or have you seen some companies crash and burn?
[34:39] How do you see the tooling ecosystem right now? And how do you see it in a few years?
[38:24] And so how do you balance that when a lot of these tools have a lot of like, bleed and overlap? And so what does that look like?
[42:40] Have you stumbled across organizations wanting to adopt AI without having the foundations such as data?
[45:28] How can we convince human curators to do machine learning?
[47:23] What are the three biggest challenges you've faced when monetizing the value of ML products. How did you overcome them?
[50:25] How do you deal with people measuring costs and values?
Machine learning has become an increasingly important means for organizations to extract value from their data. Many companies start off with successful proofs of value but face problems when scaling their capabilities afterward. By generalizing engineering problems and solving them centrally, scaling becomes much more feasible. Model serving platforms generalize the problem of turning a machine learning model in a value-generating application. Combining a serving platform with cultural shifts such as a shift-left approach enhances efficiency even further.
Bertjan is a Senior Data Engineer, with 15 years of experience in the software industry, specializing in data science and engineering for the last 10 years. He built a variety of data products and machine learning platforms. He have worked on both traditional desktop applications as well as cloud native applications in DevOps teams. He's a craftsman with a passion for delivering value through high quality software, aligning stakeholders and coaching junior and medior team members.
Axel has a background in data science. While getting his data science master degree, he did software engineering and data science projects for a wide range of customers. This experience taught him that the main complexity of data science projects lies in the software built around the predictive models. After finishing his degree, he joined BigData Republic. Axel currently helps companies bring their data science capabilities to the next level. His main interest lies in tooling that speeds up the development of machine learning applications.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Bertjan on LinkedIn: https://www.linkedin.com/in/bertjanbroeksema/
Connect with Axel on LinkedIn: https://www.linkedin.com/in/axel-goblet-5325327a/
David & Elle talk about how one of the staples of DevOps, the practice of continuous integration, can work for machine learning. Continuous integration is a tried-and-true method for speeding up development cycles and rapidly releasing software- an area where data science and ML could use some help. Making continuous integration work for ML has been challenging in the past, and we chat about new open-source tools and approaches in the Git ecosystem for leveling up development processes with big models and datasets.
|| Highlights ||
What is continuous integration and why should ML/data science teams know about it?
Why ML projects tend to fall short of DevOps best practices, like frequent check-ins and testing
How we're dealing with obstacles to get continuous integration working for ML
Also, some fun chat about how data science roles are changing and how MLOps skills fit into the data science toolkit!
The DevOps Handbook: https://amzn.to/2XH7tIT
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Elle on LinkedIn: https://www.linkedin.com/in/elle-o-brien-2a4586100/
Airflow is a renowned tool for data engineering. It helps with orchestrating ETL workloads and it's well regarded amongst machine learning engineers as well. So, how does Airflow work and how is it applied to MLOps?
In this episode, Demetrios and David are joined by Simon Darr, a Managing Consultant at Servian, with many years of experience using Airflow, along with a Byron Allen, a Senior Consultant at Servian, specializing in ML. The group discusses how Airflow works, its pros, and cons for MLOps and how it is used in practice along with a short demo.
|| Links Referenced in the Show ||
Maxime Beauchemin on Medium https://medium.com/@maximebeauchemin
The Rise of the Data Engineer: https://www.freecodecamp.org/news/the-rise-of-the-data-engineer-91be18f1e603/
Using Airflow with Kubernetes at Benevolent AI: https://www.benevolent.com/engineering-blog/using-airflow-with-kubernetes-at-benevolentai
|| Sponsored Content ||
Servian is a global data consultancy, providing advisory and delivery for data engineering and ML/AI projects. Accelerate ML is their framework to streamline and maximize the impact of ML workflows on an organization. As a part of that framework, they have a free tool used to help clients understand ML maturity. Check out the framework here along with the ML maturity assessment.
Accelerate ML framework: https://www.servian.com/accelerate-ml/
ML Maturity Assessment: https://forms.gle/4ZN9htWjSUsSBkfd7
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Simon on LinkedIn: https://www.linkedin.com/in/sdarr/
Connect with Byron on LinkedIn: https://www.linkedin.com/in/byronaallen/
Most MLOps discussion traditionally focuses on model deployment, containerization, model serving - but where do the inputs come from and where do the outputs get used? In this session we demystify parts of the data science process used to create the all-important target variable and design machine learning experiments.
We discuss some probability and statistical concepts which are useful for MLOps professionals. Knowledge of these concepts may assist practitioners working closely with data scientists or those who aspire to build complex experimentation frameworks.
Danny is a recovering data scientist who has moved over to the dark side of ML engineering in the past 2 years. He has spent multiple years deploying ML models and designing customer experiments in retail and banking sectors. Danny's passion is to guide businesses and individuals on their AI & machine learning journey. He believes a clear understanding of data strategy and applied machine learning will be a key differentiator in this brave new world. He currently provides personalised mentorship for 400+ aspiring data professionals through the #DataWithDanny community.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Danny on LinkedIn: https://www.linkedin.com/in/dannykcma/
We asked what you wanted to hear next on our Coffee sessions and the vote was in favor of feature stores! Today the usual suspects Demetrios Brinkmann and David Aponte sat down to talk with Jim Dowling CEO of Logical Clocks and Venkata Pingali CEO of scribble data to talk about feature stores, what they are, why we need them, some business implications and everything in between!
As always if you enjoyed the session let us know or reach out to us in slack!
Check out what Jim is doing around hopsworks and open sourced feature stores at Logical Clocks: https://www.logicalclocks.com/
Find out more about the feature stores that Venkata is building at Scribble Data: https://www.scribbledata.io/
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
As more and more machine learning models are deployed into production, it is imperative we have better observability tools to monitor, troubleshoot, and explain their decisions. In this talk, Aparna Dhinakaran, Co-Founder, CPO of Arize AI (Berkeley-based startup focused on ML Observability), will discuss the state of the commonly seen ML Production Workflow and its challenges. She will focus on the lack of model observability, its impacts, and how Arize AI can help.
This talk highlights common challenges seen in models deployed in production, including model drift, data quality issues, distribution changes, outliers, and bias. The talk will also cover best practices to address these challenges and where observability and explainability can help identify model issues before they impact the business. Aparna will be sharing a demo of how the Arize AI platform can help companies validate their models performance, provide real-time performance monitoring and alerts, and automate troubleshooting of slices of model performance with explainability. The talk will cover best practices in ML Observability and how companies can build more transparency and trust around their models.
Aparna Dhinakaran is Chief Product Officer at Arize AI, a startup focused on ML Observability. She was previously an ML engineer at Uber, Apple, and Tubemogul (acquired by Adobe). During her time at Uber, she built a number of core ML Infrastructure platforms including Michaelangelo. She has a bachelors from Berkeley's Electrical Engineering and Computer Science program where she published research with Berkeley's AI Research group. She is on a leave of absence from the Computer Vision PhD program at Cornell University.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Aparna on LinkedIn: https://www.linkedin.com/in/aparnadhinakaran/
In this talk, I demonstrate an example of an ML project development and production workflows which we build on top of our proprietary core - Neu.ro - using a number of open-source and proprietary tools: Jupyter Notebooks, Tensorboard, FileBrowser, PyCharm Professional, Cookiecutter, Git, DVC, Airflow, Seldon, and Grafana. I describe how we integrate each of these tools with Neu.ro, and how we can improve these integrations.
Mariya came to MLOps from a software development background. She started her career as a Java developer in JetBrains in 2011, then gradually moved to developer advocacy for JS-based APIs. In 2019, she joined Neu.ro as a platform developer advocate and then moved to the product management position. She has been obsessed with AI and ML for many years: she finished a bunch of courses, read a lot of books, and even wrote a couple of fiction stories about AI. She believes that proper tooling and decent development and operations practices are an essential success component for ML projects, as well as they are for traditional SD.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Mariya on LinkedIn: https://www.linkedin.com/in/mariya-davydova/
It can be tricky to explain MLOps to colleagues and managers who are used to traditional software engineering and DevOps, let alone your gran. We have to answer the 'Isn't that just DevOps?' question clearly, otherwise the challenges of MLOps will continue to be underestimated (potentially by us as well as others). In this session we dive into what is new about MLOps and why current mainstream DevOps alone does not solve the problems.
Ryan Dawson is an Engineer at Seldon and author of the article 'Why is DevOps for Machine Learning so Different?' You can find that article at https://hackernoon.com/why-is-devops-for-machine-learning-so-different-384z32f1
||Show Notes||
LF AI Foundation Interactive Landscape: https://landscape.lfai.foundation/
Seldon Docs: https://docs.seldon.io/en/latest/
An awesome list of references for MLOps: https://github.com/visenger/awesome-mlops
Awesome production machine learning: https://github.com/EthicalML/awesome-production-machine-learning
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Ryan on LinkedIn https://www.linkedin.com/in/ryan-dawson-501ab9123/
Python's most popular data science libraries—pandas, numpy, and scikit-learn—were designed to run on a single computer, and in some cases, using a single processor. Whether this computer is a laptop or a server with 96 cores, your compute and memory are constrained by the size of the biggest computer you have access to.
In this course, you'll learn how to use Dask, a Python library for parallel and distributed computing, to bypass this constraint by scaling our compute and memory across multiple cores. Dask provides integrations with Python libraries like pandas, numpy, and scikit-learn so you can scale your computations without having to learn completely new libraries or significantly refactoring your code.
Daniel Gerlanc has worked as a data scientist for more than decade and written software professionally for 15 years. He spent 5 years as a quantitative analyst with two Boston hedge funds before starting Enplus Advisors. At Enplus, he works with clients on data science and custom software development with a particular focus on projects requiring expertise in both areas. He teaches data science and software development at introductory through advanced levels. He has co-authored several open source R packages, published in peer-reviewed journals, and is active in local predictive analytics groups.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Daniel on LinkedIn: https://www.linkedin.com/in/dgerlanc/
How To Monitor Machine Learning Stacks - Why Current Monitoring is Unable to Detect Serious Issues and What to Do About It with Lina Weichbrodt.
Monitoring usually focusses on the “four golden signals”: latency, errors, traffic, and saturation. Machine learning services can suffer from special types of problems that are hard to detect with these signals. The talk will introduce these problems with practical examples and suggests additional metrics that can be used to detect them.
A case study demonstrates how these new metrics work for the recommendation stacks at Zalando, one of Europe’s largest fashion retailers.
Lina has 8+ years of industry experience in developing scalable machine learning models and bringing them into production. She currently works as the Machine Learning Lead Engineer in the data science group of the German online bank DKB. She previously worked at Zalando, one of Europe’s biggest online fashion retailers, where she developed real-time, deep learning personalization models for more than 32M users.
Join our Slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on Linkedin: https://www.linkedin.com/in/dpbrinkm/
Connect with Lina on Linkedin: https://www.linkedin.com/in/lina-weichbrodt-344a066a/
How to become a better data scientist: the definite guide with Alexey Grigorev
We all know what we need to do to be good data scientists: know machine learning, be able to program, be fluent in SQL and Python. That’s enough to do our job quite well. But what does it take to be a better data scientist?
The best way to grow as a data scientist is to step out of direct responsibilities and try on the hats of a product manager as well as a DevOps engineer. In particular, we should:
- be pragmatic and product-oriented
- communicate more
- get into infrastructure
After listening to this talk, you will know how exactly we should do it.
Alexey lives in Berlin with his wife and son. He’s a software engineer with a focus on machine learning. He works at OLX Group as a Lead Data Scientist. Alexey is a Kaggle master and he wrote a couple of books. One of them is “Mastering Java for Data Science” and now he’s working on another one — “Machine Learning Bookcamp”.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Alexey on LinkedIn: https://www.linkedin.com/in/agrigorev/
Companies are increasingly investing in Machine Learning (ML) to deliver new customer experiences and re-invent business processes. Unfortunately, the majority of operational ML projects never make it to production. The most significant blocker is the lack of infrastructure and tooling required to build production-ready data for ML.
Kevin Stumpf has a long history of building data infrastructure for ML, first for Uber Michelangelo, and most recently as co-founder of Tecton. Kevin will share his insights on the challenges of getting ML features to production. We’ll discuss the role of the feature store in bringing DevOps-like efficiency to building ML features. Kevin will also provide an overview of Tecton, which aims to bring an enterprise-grade feature store to every company.
Kevin co-founded Tecton where he leads a world-class engineering team that is building a next-generation feature store for operational Machine Learning. Kevin and his co-founders built deep expertise in operational ML platforms while at Uber, where they created the Michelangelo platform that enabled Uber to scale from 0 to 1000's of ML-driven applications in just a few years. Prior to Uber, Kevin founded Dispatcher, with the vision to build the Uber for long-haul trucking. Kevin holds an MBA from Stanford University and a Bachelor's Degree in Computer and Management Sciences from the University of Hagen. Outside of work, Kevin is a passionate long-distance endurance athlete.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Cristiano on LinkedIn: https://www.linkedin.com/in/cristiano-breuel/
Connect with Kevin on LinkedIn: https://www.linkedin.com/in/kevinstumpf/
David Aponte and Misha sat down and talked in depth about what the ML tool paperspace can do.
Misha Kutsovsky is a Senior Machine Learning Architect at Paperspace working on the Gradient team. He has expertise in machine learning, deep learning, distributed training, and MLOps. Previously he was on Microsoft's Windows Active Defense team building fileless malware detection software and tooling machine learning systems for Microsoft DevOps & Data Scientist teams. He holds B.S. and M.S. degrees in Electrical & Computer Engineering from Carnegie Mellon University.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter: @mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Cristiano on LinkedIn: https://www.linkedin.com/in/cristiano-breuel/
Connect with Paperspace on LinkedIn: https://www.linkedin.com/company/paperspace/
Running a Fintech on Machine Learning
For this meetup we sat down with Caique Lima and Cristiano Breuel Machine Learning Engineers at the Brasilian Fintech Nubank.
Nubank is a Fintech providing credit and banking services to more than 20 million customers. Data science has been one of the company's pillars since the beginning, and many of its critical decisions in production are made with ML, in areas such as Credit, Fraud and Customer Service. We discussed how they develop, deploy and monitor ML models, and also talk about how they built those in house solutions over the years. Today they use MLOps to support a team of more than 70 Data Scientists/ Machine Learning Engineers.
Caique is a Machine Learning Engineer at Nubank, developing software to scale decision making, this goes from model development to monitoring. Always trying to bring good practices from Software development to Data teams. Cristiano, a Machine Learning Engineer at Nubank, works to improve the efficiency and quality of ML development. Previously ML/Data Engineer at Google, specializing in MLOps, and Software Engineer at IBM.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Cris Sterry on LinkedIn: https://www.linkedin.com/in/chrissterry/
Connect with Cristiano on LinkedIn: https://www.linkedin.com/in/cristiano-breuel/
Connect with Caique on LinkedIn: https://www.linkedin.com/in/caiquelima/
DataOps and Data version Control
MLOps.community meetup #19 with the Founder and creator of DVC.org Dmitry Petrov.
Data versioning and data management are core components of MLOps and any end-to-end AI platform. What challenges are related to data versioning and how to overcome these? What are the benefits of using Git and data codification as a foundation of data versioning? And how open data versioning tools can enable an open MLOps ecosystem instead of closed end-to-end ML platforms.
DVC and other tools:
Basic modeling scenarios Automation of modeling Model deployments: to server or docker.
DVC as a model registry.
CI/CD for ML
Dmitry is a creator of open-source tool Data Version Control - DVC.org - or Git for data. He is a former data scientist at Microsoft with Ph.D. in Computer Science. Now Dmitry is working on tools for machine learning and data versioning as a Co-Founder and CEO of Iterative.AI in San Francisco, CA.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Dmitry on LinkedIn: https://www.linkedin.com/in/dmitryleopetrov/
MLOps coffee sessions coming at you with our primer episode talking bout kfserving! David Aponte and Demetrios Brinkmann dive deep into what model serving is in machine learning, what different types of serving there is, what serverless means, API endpoints, streaming and batch data and a bit of coffee vs tea banter.
||Show Notes||
ML in Production is Hard Blog article by Nikki: http://veekaybee.github.io/2020/06/09/ml-in-prod/?utm_campaign=Data_Elixir&utm_source=Data_Elixir_289
Interactive learning platform Katacoda: https://www.katacoda.com/
Github repo used in video: https://github.com/aponte411/demos
Blog on different ways to handle model serving: http://bugra.github.io/posts/2020/5/25/how-to-serve-model/
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
MLOps.community meetup #17 a deep dive into the open source ML framework Hermoine built on top of MLflow with Neylson Crepalde
Key takeaways for attendees: MLOps problems are dealt with tools but also with processes Open-source framework Hermione can help in a lot of parts of the operations process
Abstract:
In Neylson's experience with Machine Learning projects, he has encountered a series of challenges regarding agile processes to build and deploy ML models in a professional cooperative environment that fosters teamwork. While on this journey, Neylson and his team developed some of our their own solutions for these challenges. Out of this was the open-source project Hermoine born . Hermoine is a collection of solutions for these specific MLOps problems that were packaged into a library, an ML project structure framework called Hermione.
In this meetup we talk about these challenges, what they did to overcome them and how Hermione helped address these different issues along the way. We will also do a demo on how to build an ML project with Hermione.
Check out Hermoine here: https://github.com/a3data/hermione
Neylson Crepalde is a partner and MLOps Tech Lead at A3Data. He holds a PhD in Economic Sociology, a masters in Sociology of Culture, an MBA in Cultural Management and a bachelor degree in Music/Conducting. He is professor of Machine Learning and Head of Data Science Department at Izabela Hendrix Methodist Technological University. His main research interests are Machine Learning processes, Politics and Deliberation, Economic Sociology and Sociology of Education. In his PhD he has worked with Multilevel Social Network Analysis and Exponential Random Graph Models to understand the social construction of quality in an orchestras’ market.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Neyslon on LinkedIn: https://www.linkedin.com/in/neylsoncrepalde/
Venture Capital in Machine Learning Startups With John Spindler CEO of Capital Enterprise.
John Spindler CEO of Capital Enterprise. We talked about what trends he has been seeing within MLOps, ML companies and also how he evaluates a deal.
John Spindler has over 15 years experience as an entrepreneur and business advisor/consultant and as well as being responsible for the day to day management of Capital Enterprise he is also a general partner at AI Seed, an early-stage fund that invests in highly talented AI-first companies.
John is on a mission to make it possible for someone moderately intelligent, with a good idea, ambition and passion to make it as an entrepreneur.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with John on LinkedIn: https://www.linkedin.com/in/johnbspindler/
Human In The Loop Machine Learning and how to scale it with Robert Munro.
This conversation centered around the components of Human-in-the-Loop Machine Learning systems and the challenges when scaling them. Most machine learning applications learn from human examples. For example, autonomous vehicles know what a pedestrian looks like because people have spent 1000s of hours labeling “pedestrians” in videos; your smart device understands you because people have spent 1000s of hours labeling the intent of speech recordings; and machine translation services work because they are trained on 1000s of sentences that have been manually translated between languages. If you have a machine learning system that is learning from human-feedback in real-time, then there are many components to support and scale, from the machine learning models to the human interfaces and the processes for quality control.
Robert Munro is an expert in combining Human and Machine Intelligence, working with Machine Learning approaches to Text, Speech, Image and Video Processing. Robert has founded several AI companies, building some of the top teams in Artificial Intelligence. He has worked in many diverse environments, from Sierra Leone, Haiti and the Amazon, to London, Sydney and Silicon Valley, in organizations ranging from startups to the United Nations. He has shipped Machine Learning Products at startups and at/with Amazon, Google, IBM & Microsoft.
Robert has published more than 50 papers on Artificial Intelligence and is a regular speaker about technology in an increasingly connected world. He has a PhD from Stanford University. Robert is the author of Human-in-the-Loop Machine Learning (Manning Publications, 2020)
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Robert on LinkedIn: https://www.linkedin.com/in/robertjmunro/
Robert's book on Human in the loop Machine Learning:https://www.manning.com/books/human-in-the-loop-machine-learning
Blog Post "Active Learning with Pytorch": https://medium.com/pytorch/https-medium-com-robert-munro-active-learning-with-pytorch-2f3ee8ebec
The amazing Byron Allen talks to us about why MLflow and Kubeflow are not playing the same game!
ML flow vs Kubeflow is more like comparing apples to oranges or as he likes to make the analogy they are both cheese but one is an all-rounder and the other a high-class delicacy. This can be quite deceiving when analyzing the two. We do a deep dive into the functionalities of both and the pros/cons they have to offer.
Byron is a Senior Consultant at Servian - a data consultancy in Australia that also has a footprint across APAC as well as the UK. Byron is based in the London office where he helps organizations discover and build competitive advantage through their data. His focus is on client advisory and consulting delivery related to Experiments and ProductionML (i.e. data science, experimental design, ML model development, MLOps).
Byron has written about a wide range of topics including the divide between data engineer and scientist, the role of ML in the post-covid world, and Kubeflow vs. MLflow. Check it all out here: https://medium.com/@byron.allen
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Byron on LinkedIn: https://www.linkedin.com/in/byronaallen/
Resume building and Interviewing tips for data scientists and Machine learning engineers.
When on the job hunt there are some tested tips and tricks that can be applied to your resume and interviews which will give you a leg up on the rest of the competition. Anthony Kelly host of the AI in Action podcast and Executive Search Consultant focused on Machine Learning and Data Science sat down with us to talk about what some of the best resumes and CV's have in common.
We spoke about optimizing your CV/resume and maximizing opportunity once you land an interview so you can have the most amount. of options to choose from while you are on the market.
Anthony Kelly is a tech recruiter from Dublin, Ireland. Currently, he is the Country Manager for Alldus an international AI and Data Science Recruitment company. He has been working in recruitment since March 2015 and since joining the recruitment industry he has been an international top recruiter due to his performance.
Along with the above he is also the founder of the Berlin AI in Action meetup which has over 2,000 members, in addition to the meetup he also has a podcast series called AI in Action and is a Co-Founder of the Berlin AI community Awards.
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Anthony Kelly on LinkedIn: https://www.linkedin.com/in/anthonypierrekelly/
Check out the AI in Action podcast: https://podtail.com/podcast/ai-in-action-podcast/
MLOps meetup #12 // What are the advantages for a data scientist to know data engineering?
What good is learning Data Engineering skills? These days full stack is overflowing with all the different things you need to know about so why learn data Engineering now? Our guest on this meetup will make the case for what the advantages are if you do decide to learn data engineering and also go into depth on how to do data engineering in the cloud.
Dan Sullivan is a software architect and data scientist with extensive experience in big data, machine learning, data architecture, security, stream processing, and cloud architecture. Dan is the author of the official Google Cloud study guides for the Professional Architect, Professional Data Engineer, and Associate Cloud Engineer exam guides as well as NoSQL for Mere Mortals.
He is also the author of over ten LinkedIn Learning courses on data science, machine learning, SQL, data architecture, and NoSQL. He holds a Ph.D. in genetics, bioinformatics, and computational biology.
Get a copy of his new book here: https://www.wiley.com/en-us/Official+Google+Cloud+Certified+Professional+Data+Engineer+Study+Guide-p-9781119618454
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Dan Sullivan on LinkedIn: https://www.linkedin.com/in/dansullivanpdx/
MLOps community meetup #11 Machine Learning at scale in Mercado Libre with Carlos de la Torre
Mercado Libre hosts the largest online commerce and payments ecosystem in Latin America. The IT department built Fury: a PaaS framework for the development and deployment of multi-cloud, multi-technology, microservices. This platform leveraged the growth of the IT area, which now counts ~4000 people.
As such, it lacked support for machine-learning based solutions: an experimentation environment for data-scientists, infrastructure and data access support for ETL and model’s training tasks, etc. Therefore, for over a year now, they have been developing Fury Data Apps (FDA). An extension of Fury to for the design, experimentation, development and deployment of machine-learning based solutions. It is already supporting ~500 users and some high-performance production APIs.
In this meetup we talk about the main features of the platform, the supporting technology and why Carlos never accepted my Linkedin request.
Link to the article Carlos references: https://martinfowler.com/articles/cd4ml.html
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
follow us on twitter:@mlopscommunity
Sign up for the next meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Carlos on LinkedIn: https://www.linkedin.com/in/carlosdelatorre/
Follow Carlos on Twitter: @py_litox
MLOps.community meetup #9 with Charles Martin - 10 years deploying Machine Learning in the Enterprise: The Inside Scoop!
Why do some machine learning projects succeed while others fall down completely? In this discussion, we will discuss the real-world challenges that Enterprises face in deploying ML solutions, focussing on challenges with existing, legacy dev-ops environments and how certain patterns of success emerge to help combat failure.
Dr. Martin runs a boutique consultancy in San Francisco, California that supports organizations looking to research, build, and deploy data science, machine learning, and AI products. He has worked with clients like eBay, Blackrock, GoDaddy as well as widely successful startups such as Aardvark (acquired by Google) and Demand Media (the first public Billion dollar IPO after Google). He is a world-renowned researcher, collaborating with UC Berkeley on the WeightWatcher project, and has taught at UC Berkeley and Stanford, and spoken at KDD, ICML, etc. He is also currently a scientific advisor to the Page family’s Anthropocene Institute, consulting on areas including modern nuclear and quantum technologies and their response to the current pandemic.
Read more from Charles: http://calculatedcontent.com/
Join our slack community: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Charles on LinkedIn: https://www.linkedin.com/in/charlesmartin14/
Meet up #10 Saurav Chakravorty sat down with us to talk about his vision of how MLOps reflect the old Indian story of blind men and an Elephant. As a lead data scientist at Brillo Saurav has build many MLOps pipelines and experienced using different ML platforms. He comes to talk with us about the difficulties of taking an ML platform from infancy to production and other key factors he has seen within the MLOps space.
Today data science is a field that is an aggregation of people from various backgrounds - econometrics, statistics, engineering, business analysts, and data engineers. Each of these groups has different expectations from a Machine Learning platforms. But, each group faces problems that have some common challenges - improving reproducibility, reducing technical debt, reducing the time to try new experiments. The challenge before any MLOps system is to create platforms and processes that address the needs of each of these groups.
Saurav is a tinkerer in the Machine Learning world with experience in the design and development of ML applications and processes. In the past few years, he has been focused on improving the processes and tools around the Machine Learning teams. he explores the ideas of Auto ML, ML Ops, and model evaluation. He helps customers adopt and use the best tools and processes that allow them to scale their Data Science or Machine Learning tools. He has development experience in the open stack ML platforms and of late the managed ML services from Azure and AWS.
You can read his article about creating your own MLOps pipeline with open source tools here: https://towardsdatascience.com/mlops-reducing-the-technical-debt-of-machine-learning-dac528ef39de
Join our MLOps community slack:https://tinyurl.com/y75xmt7q
Come to our next MLOps meetup: https://tinyurl.com/yajmywre
Connect with Saurav on LinkedIn: https://www.linkedin.com/in/sauravchakravorty/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Linkedin, Spotify, Volvo, JP Morgan, and many other market leaders are leveraging Kubeflow to simplify the creation and the efficient deployment of Machine Learning models on Kubernetes. This presentation will provide an update on the Kubeflow 1.0 release, and review the Community’s best practices to support Critical User Journeys, which optimize ML workflows.
As a data scientist will often need to build (and save) hundreds of variants of their model, this session will provide a deeper dive into how an integrated storage solution simplifies model-building and increases ML productivity. The presentation will examine how to optimize the daily workflows of data scientists, and eliminate complex and time-consuming manual tasks. The talk will also highlight how efficient Kubeflow operations rely on Kubernetes storage primitives, such as Dynamic Volume Provisioning, Persistent Volumes and StatefulSets. This integrated solution simplifies the configuration, operations and data protection for Kubeflow and generic K8s stateful apps in production-grade, multi-user environments.
Bio:
Josh Bottum is a Kubeflow Community Product Manager. His Community responsibilities include assisting users to quantify Kubeflow business value, develop critical user journeys (CUJs), triage incoming user issues, prioritize feature delivery, write release announcements and deliver Kubeflow presentations and demonstrations.
Mr. Bottum is also a VP of Arrikto. Arrikto simplifies storage operations for stateful Kubernetes applications by enabling efficient local storage architectures with data durability and portability. Arrikto is a core code contributor to Kubeflow.
Join our MLOps community slack
Connect with Demetrios Brinkmann on LinkedIn
What does the MLOps pipeline at London Based FinTech startup TrueLayer look like?
London Based Fintech start-up TrueLayer decided to use Machine Learning instead of a rule-based system in mid-2019 and in our 7th meetup we spoke to their lead data scientist Alex Spanos about everything that entailed. During the meetup, we dove into how TrueLayer architected their MLOps pipeline for their Open Banking API: more specifically which tools they use and why, what prompted them to use machine learning, and how Alex sees the role of a Machine Learning Engineer. Alex has led the hiring process of Machine Learning Engineers and shared learnings on candidates and businesses alike.
Alex is the Lead Data Scientist at TrueLayer, focussing on building Open Banking API products powered by data. Prior to TrueLayer, he built predictive models in Financial Services, used social data to predict the “next-big-thing” in Fast Moving Consumer Goods and introduced Machine Learning techniques in subsurface imaging.
His academic background is in Applied Mathematics & Statistics.
Check out his blog entries for more info:
https://alexiospanos.com/hiring-machine-learning-engineers-part-1/
https://alexiospanos.com/hiring-machine-learning-engineers-part-2/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Alex on Linkedin: https://www.linkedin.com/in/alexspanos/
Join us on slack: https://join.slack.com/t/mlops-community/shared_invite/zt-391hcpnl-aSwNf_X5RyYSh40MiRe9Lw
In our 6th meetup, we spoke with the CEO of Scribble Data Dr. Venkata Pingali.
Scribble helps build and operate production feature engineering platforms for sub-fortune 1000 firms. The output of the platforms is consumed by data science and analytical teams. In this talk we discuss how we understand the problem space, and the architecture of the platform that we built for preparing trusted model-ready datasets that are reproducible, auditable, and quality checked, and the lessons learned in the process. We will touch upon topics like classes of consumers, disciplined data transformation code, metadata and lineage, state management, and namespaces. This system and discussion complements work done on data science platforms such as Domino and Dotscience.
Bio: Dr. Venkata Pingali is Co-Founder and CEO of Scribble Data, an ML Engineering company with offices in India and Canada. Scribble’s flagship enterprise product, Enrich, enables organizations to address 10x analytics/data science usecases through trusted production datasets. Before starting Scribble Data, Dr. Pingali was VP of Analytics at a data consulting firm and CEO of an energy analytics firm. He has a BTech from IIT Mumbai and a PhD from USC in Computer Science.
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Venkata on LinkedIn: https://www.linkedin.com/in/pingali/
In our 5th meetup, we spoke with the Brasilian ML Engineer Flavio Clesio.
Machine Learning Systems play a huge role in several businesses from the Banking industry to recommender systems in entertainment applications until health domains. The era of "A Data Scientist with a Script in a single machine" is officially over in high stakes ML.
We're entering an era of Machine Learning Operations (MLOps) where those critical applications that impact society and businesses need to be aware of aspects like active failures and latent conditions. This talk will discuss risk assessment in ML Systems from the perspective of reliability, safety and especially causal aspects that can lead to the rise of silent risks in said systems.
Slides to the talk can be found here
Bio:
Flavio Clesio is Machine Learning Engineer (NLP, CV, Marketplace RecSys) and at the moment works at MyHammer AG, where he helps build Core Machine Learning applications to exploit revenue opportunities and automation in decision making.
Prior to MyHammer, Flavio was a Data Intelligence lead in the mobile industry, and business intelligence analyst in financial markets, specifically in Non-Performing Loans. He holds a master’s degree in computational intelligence applied in financial markets (exotic credit derivatives).
This was a virtual fireside chat between Flavio Clesio, Demetrios Brinkmann and the MLOps community. Relevant links can be found below. Join our MLOps slack community: https://bit.ly/3aOTwgR and register for the next meetup here.
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with FLavio Clesio on Linkedin: https://www.linkedin.com/in/flavioclesio/
MLOps Community Meetup #4 With Shubhi Jain
In the 4th online meetup for our MLOps.community We spoke with Shubhi Jain, Machine Learning Engineer and an all-around great guy!
Every organization is leveraging machine learning (ML) to provide increasing value to their customers and understand their business. You may have created models too. But, how do you scale this process now? In this case study, we looked at how to pinpoint inefficiencies in your ML data flow, how SurveyMonkey tackled this, and how to make your data more usable to accelerate ML model development.
Shubhi Jain is a machine learning engineer at SurveyMonkey where he develops and implements machine learning systems for its products and teams. Occasionally, he’ll create YouTube videos about Machine Learning in collaboration with Springboard, an e-learning platform. He’s always excited to bring his expertise and passion for Data and AI systems to the rest of the industry. In his free time, Shubhi likes hiking with his dog and accelerating his hearing loss at live music shows.
This was a virtual fireside chat between Shubhi Jain, Demetrios Brinkmann and the MLOps community. Relevant links can be found below. Join our MLOps slack community: https://bit.ly/3aOTwgR and register for the next meetup here.
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Shubhi Jain on Linkedin: https://www.linkedin.com/in/shubhankarjain/
Check out more of Shubhi on youtube:
MLOps community meetup #3! Last Wednesday we talked to Phil Winder, CEO, Winder Research.
//Abstract
Phil Winder of Winder Research joined us for the 3rd instalment of our MLOps community meetup. In this clip taken from the long conversation, he speaks about why or why not he sees companies automating the retraining of Machine Learning Models. You can find the whole conversation here: https://www.youtube.com/watch?v=MRES5IxVnME
The topic of conversation for our virtual meetup was an in-depth look at a pyramid of software engineering best practices that built up to incorporate data science best practices. That is to say, we analyzed “the essentials”, "nice to have" and "optimal" ways of doing data science.
Machine Learning/Data Science/AI is an extension of the technical stack. So you can't really talk about Data science best practices without accidentally talking about software engineering best practices. For example, model provenance doesn't count for anything if you don't have code or container provenance. Just as Maslow has the basic human needs so too do we have basic MLOps needs. Where does "MLOps", as a "thing", starts and end? For example, the four very reasonable best practices of the operation of models, but these are usually consumed into higher-level abstractions because there is a lot more to do than "just" provenance.
//Bio
Dr Phil Winder is a multidisciplinary software engineer and data scientist. As the CEO of Winder Research, a Cloud-Native data science consultancy, he helps startups and enterprises improve their data-based processes, platforms, and products. Phil specializes in implementing production-grade cloud-native machine learning and was an early champion of the MLOps movement. More recently, Phil has authored a book on Reinforcement Learning (RL) (https://rl-book.com) which provides an in-depth introduction of industrial RL to engineers. He has thrilled thousands of engineers with his data science training courses in public, private, and on the O’Reilly online learning platform. Phil’s courses focus on using data science in industry and cover a wide range of hot yet practical topics, from cleaning data to deep reinforcement learning. He is a regular speaker and is active in the data science community. Phil holds a PhD and M.Eng. in electronic engineering from the University of Hull and lives in Yorkshire, U.K., with his brewing equipment and family.
//This was a virtual fireside chat between Phil Winder and Demetrios Brinkmann. relevant links can be found below:
Join our MLOps slack community: https://bit.ly/3aOTwgR
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Phil on LinkedIn: https://www.linkedin.com/in/drphilwinder/
Follow Phil on Twitter: https://twitter.com/DrPhilWinder
Learn more about Phil's company Winder research: https://winderresearch.com/
For this episode, we are joined by Charles Radclyffe, who until very recently was the Head of AI at Fidelity. Some of his other feats include starting 3 companies, TEDx talks, and advising the likes of HSBC, Barclays, Morgan & Stanley, and Deutsche Bank. He has focused his career on solving tough technology challenges for some of the world's largest organizations, for more on him, follow on twitter or connect on LinkedIn
Governance is coming for us all, but it’s especially pertinent in regulated industries such as the finance sector. Financial institutions must be mindful of how their machine learning models are being used and experimented with as regulators are keen to understand the quality of controls across the industry.
Our conversation is centered around Charles’ past experiences heading up the AI capability for a large organization in the financial industry, and his learnings during that time. We will also touch on what ideal AI/ML governance looks like in his eyes and where he sees we need to focus our attention for future success within this area. What do data scientists and ML engineers need to learn about governance to ensure business success as laws are continually changing?
This episode is a virtual fireside chat for the first 40 minutes and in the last 20 minutes we open up the floor to any questions. Please feel free to join our slack channel or forum to chat more about MLOps.
Link to MIT Techlash blog Charles wrote.
Join our open community where we discuss everything MLOps: https://mlops.community/ Join our MLOps slack channel: https://bit.ly/33wDUf1 MLOps.community forums: https://forum.mlops.community/ Sign up for the next weekly meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
The 1st MLOps.community meetup on 3.18.2020 featuring Luke Marsden from Dotscience.
What is MLOps and how can it help me work remotely? The first episode of our weekly MLOps community virtual meetup with CEO and founder of the MLOps platform dotscience Luke Marsden talk to us about the current state of Machine Learning, what some of the main difficulties are at this stage when developing models, how the machine learning lifecycle differs from traditional software development and a deep dive of collaboration for data science teams in a fully remote world.
MLOps is the intersection of three disciplines: software engineering, DevOps and machine learning. MLOps refers to the entire end-to-end lifecycle of getting models from lab to live where they can start delivering value.
What do software engineers and DevOps need to learn about machine learning to ensure that it can be integrated into their dev & deployment pipelines? What do data scientists and ML engineers need to learn about DevOps, model deployment and monitoring to ensure they can effectively deploy their work without racking up tonnes of technical debt? And now that working from home is fast becoming the new normal, how can MLOps help my team stay efficient when asynchronous collaboration is needed, something our software engineering and DevOps friends have already mastered?
MLOps is a complex discipline due to the many more moving parts involved than regular software DevOps, in this inaugural MLOps.community meetup we'll explore and navigate this new space together and give you a guide on how to avoid the most common pitfalls and challenges getting AI into production and collaborating effectively with your team – even when you're distributed.
Join our open community where we discuss everything MLOps: https://mlops.community/
Join our MLOps slack channel: https://bit.ly/33wDUf1
MLOps.community forums: https://forum.mlops.community/
Sign up for the next weekly meetup: https://zoom.us/webinar/register/WN_a_nuYR1xT86TGIB2wp9B1g
En liten tjänst av I'm With Friends. Finns även på engelska.