Gradient Dissent: Conversations on AI Podcast Por Lukas Biewald capa

Gradient Dissent: Conversations on AI

Gradient Dissent: Conversations on AI

De: Lukas Biewald
Ouça grátis

Sobre este título

Join Lukas Biewald on Gradient Dissent, an AI-focused podcast brought to you by Weights & Biases. Dive into fascinating conversations with industry giants from NVIDIA, Meta, Google, Lyft, OpenAI, and more. Explore the cutting-edge of AI and learn the intricacies of bringing models into production.All rights reserved Economia
Episódios
  • The CEO Behind the Fastest-Growing AI Inference Company | Tuhin Srivastava
    Nov 18 2025

    In this episode of Gradient Dissent, Lukas Biewald talks with Tuhin Srivastava, CEO and founder of Baseten, one of the fastest-growing companies in the AI inference ecosystem. Tuhin shares the real story behind Baseten’s rise and how the market finally aligned with the infrastructure they’d spent years building.

    They get into the core challenges of modern inference, including why dedicated deployments matter, how runtime and infrastructure bottlenecks stack up, and what makes serving large models fundamentally different from smaller ones.


    Tuhin also explains how vLLM, TensorRT-LLM, and SGLang differ in practice, what it takes to tune workloads for new chips like the B200, and why reliability becomes harder as systems scale.


    The conversation dives into company-building, from killing product lines to avoiding premature scaling while navigating a market that shifts every few weeks.


    Connect with us here:

    Tuhin Srivastva: https://www.linkedin.com/in/tuhin-srivastava/

    Lukas Biewald: https://www.linkedin.com/in/lbiewald/

    Weights & Biases: https://www.linkedin.com/company/wandb/

    Exibir mais Exibir menos
    59 minutos
  • The Startup Powering The Data Behind AGI
    Sep 16 2025

    In this episode of Gradient Dissent, Lukas Biewald talks with the CEO & founder of Surge AI, the billion-dollar company quietly powering the next generation of frontier LLMs. They discuss Surge's origin story, why traditional data labeling is broken, and how their research-focused approach is reshaping how models are trained.

    You’ll hear why inter-annotator agreement fails in high-complexity tasks like poetry and math, why synthetic data is often overrated, and how Surge builds rich RL environments to stress-test agentic reasoning. They also go deep on what kinds of data will be critical to future progress in AI—from scientific discovery to multimodal reasoning and personalized alignment.


    It’s a rare, behind-the-scenes look into the world of high-quality data generation at scale—straight from the team most frontier labs trust to get it right.


    Timestamps:

    00:00 – Intro: Who is Edwin Chen?

    03:40 – The problem with early data labeling systems

    06:20 – Search ranking, clickbait, and product principles

    10:05 – Why Surge focused on high-skill, high-quality labeling

    13:50 – From Craigslist workers to a billion-dollar business

    16:40 – Scaling without funding and avoiding Silicon Valley status games

    21:15 – Why most human data platforms lack real tech

    25:05 – Detecting cheaters, liars, and low-quality labelers

    28:30 – Why inter-annotator agreement is a flawed metric

    32:15 – What makes a great poem? Not checkboxes

    36:40 – Measuring subjective quality rigorously

    40:00 – What types of data are becoming more important

    44:15 – Scientific collaboration and frontier research data

    47:00 – Multimodal data, Argentinian coding, and hyper-specificity

    50:10 – What's wrong with LMSYS and benchmark hacking

    53:20 – Personalization and taste in model behavior

    56:00 – Synthetic data vs. high-quality human data


    Follow Weights & Biases:

    https://twitter.com/weights_biases

    https://www.linkedin.com/company/wandb

    Exibir mais Exibir menos
    56 minutos
  • Arvind Jain on Building Glean and the Future of Enterprise AI
    Aug 5 2025

    In this episode of Gradient Dissent, Lukas Biewald sits down with Arvind Jain, CEO and founder of Glean. They discuss Glean's evolution from solving enterprise search to building agentic AI tools that understand internal knowledge and workflows. Arvind shares how his early use of transformer models in 2019 laid the foundation for Glean’s success, well before the term "generative AI" was mainstream.

    They explore the technical and organizational challenges behind enterprise LLMs—including security, hallucination suppression—and when it makes sense to fine-tune models. Arvind also reflects on his previous startup Rubrik and explains how Glean’s AI platform aims to reshape how teams operate, from personalized agents to ever-fresh internal documentation.

    Follow Arvind Jain: https://x.com/jainarvind

    Follow Weights & Biases: https://x.com/weights_biases


    Timestamps:

    [00:01:00] What Glean is and how it works

    [00:02:39] Starting Glean before the LLM boom

    [00:04:10] Using transformers early in enterprise search

    [00:06:48] Semantic search vs. generative answers

    [00:08:13] When to fine-tune vs. use out-of-box models

    [00:12:38] The value of small, purpose-trained models

    [00:13:04] Enterprise security and embedding risks

    [00:16:31] Lessons from Rubrik and starting Glean

    [00:19:31] The contrarian bet on enterprise search

    [00:22:57] Culture and lessons learned from Google

    [00:25:13] Everyone will have their own AI-powered "team"

    [00:28:43] Using AI to keep documentation evergreen

    [00:31:22] AI-generated churn and risk analysis

    [00:33:55] Measuring model improvement with golden sets

    [00:36:05] Suppressing hallucinations with citations

    [00:39:22] Agents that can ping humans for help

    [00:40:41] AI as a force multiplier, not a replacement

    [00:42:26] The enduring value of hard work



    Exibir mais Exibir menos
    44 minutos
Ainda não há avaliações