Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

The podcast by and for AI Engineers! In 2023, over 1 million visitors came to Latent Space to hear about news, papers and interviews in Software 3.0. We cover Foundation Models changing every domain in Code Generation, Multimodality, AI Agents, GPU Infra and more, directly from the founders, builders, and thinkers involved in pushing the cutting edge. Striving to give you both the definitive take on the Current Thing down to the first introduction to the tech you'll be using in the next 3 months! We break news and exclusive interviews from OpenAI, tiny (George Hotz), Databricks/MosaicML (Jon Frankle), Modular (Chris Lattner), Answer.ai (Jeremy Howard), et al. Full show notes always on https://latent.space www.latent.space

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Tue Jul 23 2024

Llama 2, 3 & 4: Synthetic Data, RLHF, Agents on the path to Open Source AGI

Natural Language ProcessingModel DevelopmentTraining MethodsAI ModelsDeep Learning Technology

The episode covers various topics including the background and journey into NLP, development of annotation projects and model size considerations, closing the gap with GPT-4 and training improvements in Lama 3, multilingual capabilities and synthetic data generation in Lama 3, Lama 3 training and model improvement, supervised fine-tuning, reinforcement learning with human feedback, and the teacher-critic method, advancements in AI models and evaluating progress, state-of-the-art results and integration of world models, thinking in latent space and balancing research with product needs, rapid evolution of deep learning technology, and common sense thinking.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Fri Jul 12 2024

Benchmarks 201: Why Leaderboards > Arenas >> LLM-as-Judge

Hugging FaceOpenLLM leaderboardautomated benchmarkshuman evaluationsmulti-turn conversations

The episode covers topics such as the background and journey to Hugging Face, the OpenLLM leaderboard and community engagement, automated benchmarks and human evaluations, multi-turn conversations and unique evaluation methods, multi-choice evaluations and commercial considerations, thorough evaluation process and future benchmark areas.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Fri Jul 05 2024

The 10,000x Yolo Researcher Metagame — with Yi Tay of Reka

AI engineeringML researchNode failuresMultimodal modelsEfficiency papers

This episode covers recent events at the World's Fair, the evolution of ML research, the challenges of node failures in large model training runs, architectural experiments, trends in multimodal models, efficiency papers, and insights on the open-source community and productivity practices. It also explores the approach to research, building a strong AI ecosystem, government influence on AI development, brain drain, and the role of senior individual contributors.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Tue Jun 25 2024

State of the Art: Training >70B LLMs on 10,000 H100 clusters

DatabricksMosaicInfrastructureCluster ManagementGPU Training

This episode covers recent developments at Databricks and Mosaic, infrastructure and training scripts, non-standard cluster setup, managing GPU training and custom infrastructure, diagnosing and debugging errors in machine learning infrastructure, complexity and power of modern machine learning infrastructure, scaling up data storage and processing, evaluation and optimization of models, fixing ambiguous examples and advancements in evaluations, embracing imperfection and ambiguity in deep learning, practical applications and long context utilization, interacting with structured data sources and choosing the right abstraction, balancing the use of code and determining proper application.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Tue Jun 25 2024

[High Agency] AI Engineer World's Fair Preview

AI engineeringAI engineersproduct developmentAI industryAI startups

The episode covers various aspects of AI engineering, including the AI Engineer World's Fair, qualifications and skills for AI engineers, their role in product development, the debate and evolution of the AI engineer role, the AI Engineering World Fair event, success strategies for AI startups, trends in AI product tooling, key battlegrounds and insights in the AI industry, and advancements in AI models and use cases.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Fri Jun 21 2024

How To Hire AI Engineers — with James Brady & Adam Wiggins of Elicit

AI EngineeringMachine LearningLanguage ModelsResilient SystemsML-First Mindset

This episode covers various aspects of AI engineering at Illicit, including transitioning to AI and ML, building resilient systems with language models, interviews and system design, evaluating new AI models, embracing the ML-first mindset, sourcing talent for AI engineering roles, and hiring AI engineers. It also highlights the blend of systematic control and creativity in AI engineering and the need for a forum to share insights in the field.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Tue Jun 11 2024

How AI is eating Finance — with Mike Conover of Brightwave

Artificial IntelligenceFinanceMachine LearningInvestment Strategies

This episode explores the intersection of artificial intelligence (AI) and finance. Mike Honnebour, founder of BrightWave, discusses the training of language models, the capabilities of AI in the finance industry, and the challenges and opportunities in applying AI to investment strategies. The episode covers topics such as machine learning systems, user interaction, retrieval systems, fine-tuning language models, financial models, AI investments, and the shelf life of models.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Mon Jun 10 2024

ICLR 2024 — Best Papers & Talks (Benchmarks, Reasoning & Agents) — ft. Graham Neubig, Aman Sanger, Moritz Hardt)

ICLRlanguage modelsbenchmarkingweb navigationcoding agents

The podcast covers the 12th International Conference on Learning Representations (ICLR) in Vienna, Austria. Part one discussed the best papers of ICLR in various sections like image generation, computer vision, and attention algorithms. Part two focuses on papers related to LLM reasoning and agents in four sections: latent space chat with Graham Newbig, benchmarking issues, agent building blocks, and proposed agent systems from Google DeepMind. The podcast also mentions an upcoming AI engineering event in June with various speakers from big companies and startups discussing AI strategy and leadership.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Thu May 30 2024

How to train a Million Context LLM — with Mark Huang of Gradient.ai

AIMachine LearningModel TrainingContext LengthModel Quality

This episode covers various topics related to AI and machine learning. Mark Wang from Gradient discusses his transition from finance to tech and the founding story of Gradient. Key insights include the focus on out-of-domain problems in machine learning and out-of-domain generalization in AI. Attention mechanisms, positional encodings, and extending context length in models are also explored. The episode delves into Gemini's million token context, recent papers on model training, manipulating models, benchmarking, scaling token context size, Google's focus on model quality and handling evolving context, building technology with early fusion models, and staying updated on AI research. The importance of context length, theta scaling, and prioritizing valuable tasks is also discussed.

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al

Mon May 27 2024

ICLR 2024 — Best Papers & Talks (ImageGen, Vision, Transformers, State Space Models, and other Learning Representations) ft. Christian Szegedy, Ilya Sutskever, Durk Kingma

Variational AutoencodersVAEsDeep LearningProbabilistic ModelsDiffusion Models

This episode covers a wide range of topics related to Variational Autoencoders (VAEs), including their introduction and evolution, advantages and challenges of latent variable models, applications of VAEs, diffusion models and scalability of VAEs, interpreting diffusion models and concept decomposition, concept manipulation and adversarial examples in generative models, unsupervised learning and distribution matching, compression and prediction in unsupervised learning, training with compression objective and adversarial learning, adversarial examples and vulnerability of neural networks, outliers and attention maps in vision transformers, pause tokens and data selection for model training, data selection and sub-selection schemes, Dora and self-supervised learning methods, efficient training and inference in large language models, adaptive KV cache compression and efficient training techniques, efficient computation in large language models, state space models and efficient computation, diffusion models and state space models. The episode provides key insights into these topics and their implications.

Page 1 of 5