Episodes

  • Episode 1: Introducing Vanishing Gradients
    Feb 16 2022
    In this brief introduction, Hugo introduces the rationale behind launching a new data science podcast and gets excited about his upcoming guests: Jeremy Howard, Rachael Tatman, and Heather Nolis! Original music, bleeps, and blops by local Sydney legend PlaneFace (https://planeface.bandcamp.com/album/fishing-from-an-asteroid)!
    Show more Show less
    5 mins
  • Episode 38: The Art of Freelance AI Consulting and Products: Data, Dollars, and Deliverables
    Nov 4 2024
    Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions. This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls. LINKS The livestream on YouTube (https://youtube.com/live/9CFs06UDbGI?feature=share) Jason's Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business (https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&utm_medium=partner&utm_source=instructor) Hugo's upcoming course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) Jason's website (https://jxnl.co/) Jason's indie consulting newsletter (https://indieconsulting.podia.com/) Your AI Product Needs Evals by Hamel Husain (https://hamel.dev/blog/posts/evals/) What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/) Dear Future AI Consultant by Jason (https://jxnl.co/writing/#dear-future-ai-consultant) Alex Hormozi's books (https://www.acquisition.com/books) The Burnout Society by Byung-Chul Han (https://www.sup.org/books/theory-and-philosophy/burnout-society) Jason on Twitter (https://x.com/jxnlco) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne) Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
    Show more Show less
    1 hr and 24 mins
  • Episode 37: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 2
    Oct 8 2024
    Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences. This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more. In this episode, we cover: The Prompt Report: A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques. Security Risks and Prompt Hacking: A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts. AI’s Impact Across Fields: A discussion on how generative AI is reshaping various domains, including the social sciences and security. Multimodal AI: Updates on how large language models (LLMs) are expanding to interact with images, code, and music. Case Study - Detecting Suicide Risk: A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges. The episode concludes with a reflection on the evolving landscape of LLMs and multimodal AI, and what might be on the horizon. If you haven’t yet, make sure to check out Part 1, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative. LINKS The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share) The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608) Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne) Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
    Show more Show less
    51 mins
  • Episode 36: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 1
    Sep 30 2024
    Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences. This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more. In this first part, * we’ll explore the critical role of prompt engineering, * & diving into adversarial techniques like prompt hacking and * the challenges of evaluating these techniques. * we’ll examine the impact of few-shot learning and * the groundbreaking taxonomy of prompting techniques from the Prompt Report. Along the way, * we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. * we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and * how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code. Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more. LINKS The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share) The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608) Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne) Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
    Show more Show less
    1 hr and 4 mins
  • Episode 35: Open Science at NASA -- Measuring Impact and the Future of AI
    Sep 19 2024
    Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore Measuring the Impact of Open Science: How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments. The Process of Scientific Discovery: Insights into the collaborative nature of research and how breakthroughs are achieved at NASA. ** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis. Addressing Challenges in Open Science: The complexities of implementing open science within government agencies and research environments. Reforming Incentive Systems: How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing. The Future of Open Science: How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility. This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful. LINKS The livestream on YouTube (https://youtube.com/live/VJDg3ZbkNOE?feature=share) NASA's Open Science 101 course <-- do it to learn and also to get NASA Swag! (https://openscience101.org/) Science Cast (https://sciencecast.org/) NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data (https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model) Jake VanderPlas' daily conundrum tweet from 2013 (https://x.com/jakevdp/status/408678764705378304) Replit, "an AI-powered software development & deployment platform for building, sharing, and shipping software fast." (https://replit.com/)
    Show more Show less
    58 mins
  • Episode 34: The AI Revolution Will Not Be Monopolized
    Aug 22 2024
    Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they've had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike. In this wide-ranging discussion, we dive into: • The evolution of applied NLP and its role in industry • The balance between large language models and smaller, specialized models • Human-in-the-loop distillation for creating faster, more data-private AI systems • The challenges and opportunities in NLP, including modularity, transparency, and privacy • The future of AI and software development • The potential impact of AI regulation on innovation and competition We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world. Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field. Whether you're a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field's most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications. LINKS The livestream on YouTube (https://youtube.com/live/-6o5-3cP0ik?feature=share) How S&P Global is making markets more transparent with NLP, spaCy and Prodigy (https://explosion.ai/blog/sp-global-commodities) A practical guide to human-in-the-loop distillation (https://explosion.ai/blog/human-in-the-loop-distillation) Laws of Tech: Commoditize Your Complement (https://gwern.net/complement) spaCy: Industrial-Strength Natural Language Processing (https://spacy.io/) LLMs with spaCy (https://spacy.io/usage/large-language-models) Explosion, building developer tools for AI, Machine Learning and Natural Language Processing (https://explosion.ai/) Back to our roots: Company update and future plans, by Matt and Ines (https://explosion.ai/blog/back-to-our-roots-company-update) Matt's detailed blog post: back to our roots (https://honnibal.dev/blog/back-to-our-roots) Ines on twitter (https://x.com/_inesmontani) Matt on twitter (https://x.com/honnibal) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne) Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams!
    Show more Show less
    1 hr and 43 mins
  • Episode 33: What We Learned Teaching LLMs to 1,000s of Data Scientists
    Aug 12 2024
    Hugo speaks with Dan Becker and Hamel Husain, two veterans in the world of data science, machine learning, and AI education. Collectively, they’ve worked at Google, DataRobot, Airbnb, Github (where Hamel built out the precursor to copilot and more) and they both currently work as independent LLM and Generative AI consultants. Dan and Hamel recently taught a course on fine-tuning large language models that evolved into a full-fledged conference, attracting over 2,000 participants. This experience gave them unique insights into the current state and future of AI education and application. In this episode, we dive into: * The evolution of their course from fine-tuning to a comprehensive AI conference * The unexpected challenges and insights gained from teaching LLMs to data scientists * The current state of AI tooling and accessibility compared to a decade ago * The role of playful experimentation in driving innovation in the field * Thoughts on the economic impact and ROI of generative AI in various industries * The importance of proper evaluation in machine learning projects * Future predictions for AI education and application in the next five years * We also touch on the challenges of using AI tools effectively, the potential for AI in physical world applications, and the need for a more nuanced understanding of AI capabilities in the workplace. During our conversation, Dan mentions an exciting project he's been working on, which we couldn't showcase live due to technical difficulties. However, I've included a link to a video demonstration in the show notes that you won't want to miss. In this demo, Dan showcases his innovative AI-powered 3D modeling tool that allows users to create 3D printable objects simply by describing them in natural language. LINKS The livestream on YouTube (https://youtube.com/live/hDmnwtjktsc?feature=share) Educational resources from Dan and Hamel's LLM course (https://parlance-labs.com/education/) Upwork Study Finds Employee Workloads Rising Despite Increased C-Suite Investment in Artificial Intelligence (https://investors.upwork.com/news-releases/news-release-details/upwork-study-finds-employee-workloads-rising-despite-increased-c) Episode 29: Lessons from a Year of Building with LLMs (Part 1) (https://vanishinggradients.fireside.fm/29) Episode 30: Lessons from a Year of Building with LLMs (Part 2) (https://vanishinggradients.fireside.fm/30) Dan's demo: Creating Physical Products with Generative AI (https://youtu.be/U5J5RUOuMkI?si=_7cYLYOU1iwweQeO) Build Great AI, Dan's boutique consulting firm helping clients be successful with large language models (https://buildgreat.ai/) Parlance Labs, Hamel's Practical consulting that improves your AI (https://parlance-labs.com/) Hamel on Twitter (https://x.com/HamelHusain) Dan on Twitter (https://x.com/dan_s_becker) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne)
    Show more Show less
    1 hr and 25 mins
  • Episode 32: Building Reliable and Robust ML/AI Pipelines
    Jul 27 2024
    Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook. In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We'll explore: The fascinating journey from classic machine learning to the current LLM revolution Why Shreya believes most ML problems are actually data management issues The concept of "data flywheels" for LLM applications and how to implement them The intriguing world of evaluating AI systems - who validates the validators? Shreya's work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences The importance of human-in-the-loop processes in AI development The future of low-code and no-code tools in the AI landscape We'll also touch on the potential pitfalls of over-relying on LLMs, the concept of "Habsburg AI," and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes. Whether you're a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems. LINKS The livestream on YouTube (https://youtube.com/live/hKV6xSJZkB0?feature=share) Shreya's website (https://www.sh-reya.com/) Shreya on Twitter (https://x.com/sh_reya) Data Flywheels for LLM Applications (https://www.sh-reya.com/blog/ai-engineering-flywheel/) SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines (https://arxiv.org/abs/2401.03038) What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/) Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences (https://arxiv.org/abs/2404.12272) Operationalizing Machine Learning: An Interview Study (https://arxiv.org/abs/2209.09125) Vanishing Gradients on Twitter (https://twitter.com/vanishingdata) Hugo on Twitter (https://twitter.com/hugobowne) In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild! If you want to dive deep into Shreya's work and related topics through their chats, you can check them all out here: Outerbounds' Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners (https://www.youtube.com/watch?v=7zB6ESFto_U) The Past, Present, and Future of Generative AI (https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS) LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering (https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0) Lessons from a Year of Building with LLMs (https://youtube.com/live/c0gcsprsFig?feature=share) Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams!
    Show more Show less
    1 hr and 15 mins