Kennis die lekker wegluistert

Onze podcasts - met liefde gemaakt door team Sourcelabs - en met (meer dan) een vleugje AI. Lekker voor onderweg.

Een dagelijkse AI-gegenereerde podcast over agentic AI, developer tooling en tech trends — volledig autonoom geproduceerd. Beschikbaar als RSS feed.

The Daily Agentic AI Podcast - 2026-04-10

Vercel pushed “agentic infrastructure” as the future of the cloud: deployment surfaces and long-running “token delivery” compute for agents, plus a platform vision of self-healing with human approval, backed by AI SDK 6, AI Gateway monitoring/routing, and GLM 5.1 on the gateway for long-horizon plan→execute→test loops. Anthropic’s Claude Cowork went GA with faster Claude Code file at-mentions, and new agent runtime tooling like Claude Code’s native Monitor/background streaming and pi-monitor for background Pi agent command execution; OpenAI also rebalanced ChatGPT pricing with a $100 Pro tier to enable heavier Codex use. Research emphasized moving beyond static “generate code” toward observation and profiling: DAIRA integrates dynamic analysis into an issue-resolution loop (reported gains on SWE-bench Verified with lower cost), while agent-written tests often act only as observational feedback rather than significantly improving outcomes (contrasted with TOP-style test validation). Security and multi-agent work covered PAGENT’s dynamic-guided PoC generation, LLM-based interprocedural vulnerability detection across languages, limits of library-hallucination mitigation, smart-contract auditing with coordinated agents (SPEAR), agents implemented as native POSIX processes (Quine), and persistent externalized memory/skills via tools like ByteRover and broader “externalized agent capabilities” architectures.

The Daily Agentic AI Podcast - 2026-04-09

OSGym introduced OS-level infrastructure for GUI computer-use agents by running over 1,000 parallel Dockerized OS replicas via copy-on-write disk cloning and a pre-warmed runner pool, enabling 1,024 replicas to generate 1,400+ trajectories per minute and fine-tune Qwen2.5-VL with strong OSWorld success on Verified benchmarks. Meta launched Muse Spark (hosted on meta.ai) alongside agentic tool modes (Instant/Thinking and a sub-agent “spawn” pattern), while Alibaba’s Qwen3.6 Plus added 1M-token native vision with strong benchmark value versus GPT/Claude at far lower cost, and curriculum learning discussions focused on how to stage data for gradient-free hill-climbing and how ordering/transfer across agent tasks matters. Anthropic and Vercel emphasized production substrates and compliance for long-running agentic systems: Anthropic Managed Agents target hosted, long-duration autonomy, Vercel AI Gateway’s “Fast mode” boosts Opus 4.6 token speeds for agentic coding, and team-wide ZDR plus “disallow prompt training” provides a compliance routing layer across providers. Vercel also pushed agentic microfrontend management (CLI + editor “AI skill”) and the v0 + new.website merge to support end-to-end, production-ready website lifecycles with agent-aware features like forms, DB-backed submissions, SEO, and CMS.

The Daily Agentic AI Podcast - 2026-04-08

GLM-5.1 (open-weight, MIT licensed) pushes long-horizon agentic coding with asynchronous reinforcement learning, sustaining hundreds of iterations and thousands of tool calls for up to eight hours while achieving strong SWE-Bench Pro results (58.4%). Meta also released Muse Spark, a top-ranked multimodal reasoning model with tool use and Contemplating mode, while Anthropic’s Claude Mythos Preview is restricted to security partners because it can autonomously find and chain exploits—paired with new evidence that AI-generated code is “broken by default” (55.8% vulnerable) and typical security instructions/scanners help little. Agentic security and evaluation tooling advanced alongside these model releases (Vulnsage-style exploit frameworks, AutoPT taxonomy, LangSmith/HF Agent Traces, LangChain Fleet + TryArcade MCP tools, APEX-Agents-AA), while coding-agent performance is increasingly measured by beyond-pass metrics like design-constraint compliance, with efficiency/repair improvements from Squeez/CODESTRUCT/DAIRA and Google’s Smart Paste auto-fix feature.

The Daily Agentic AI Podcast - 2026-04-07

Vercel’s monorepo added an LLM-based risk classifier with conservative LOW/HIGH gating, hard rules (e.g., many-file changes or CODEOWNERS paths), phased kill-switch rollouts, and adversarial hardening—achieving 58% auto-merges of low-risk PRs with zero reverts and much faster merge times. Research and tools also span coding-agent architectures and training efficiency (Inside the Scaffold taxonomy, STITCH fewer-but-better trajectories), empirical GitHub evidence of agent edits plus integration pain (AgenticFlict merge conflicts), and production/safety advances (LangSmith/LangChain cost monitoring, DebugHarness autonomous security patching, ABTest behavior-driven anomaly testing, SWE-EVO long-horizon evolution benchmarks), alongside model/datasight and IDE practicality (Gemini 3.1 Pro in Augment Code, SADU VLM diagram limits, Smart Paste acceptance impact). Legal risk surfaced via “Alignment Whack-a-Mole,” where fine-tuning on Murakami unlocked verbatim copyrighted novel reproduction, and interoperability/open collaboration progressed through agent trace sharing and session mirroring (pi-magic-docs/agent traces/agent-session-bridge).

The Daily Agentic AI Podcast - 2026-04-06

Netflix VOID demonstrates physics-aware video object removal by erasing both an object and the physical interactions it caused, using synthetic Blender/Kubric data, quadmask encoding, and a two-pass CogVideoX-based inference pipeline. Agentic tooling and research also featured AutoAgent (overnight meta-optimization of prompts/tools to score better benchmark runs), Karpathy’s “idea files” for sharing abstract agent specs instead of code, AlphaEvolve (LLM-evolved rewrites of game theory algorithms like CFR/PSRO to beat prior results), and AutoKernel (agentic GPU kernel optimization with commit-backed regression control). The episode further covered runtime tracing and observability for coding agents, SWE-STEPS for more realistic sequential coding evaluation (including inflated PR-only success and rising debt/complexity), behavioral variance as a key driver of agent failures, terminal-only enterprise automation vs richer orchestration (KAIJU intent-gated execution), and MCP/infrastructure updates like Waldium, Nuxt MCP tooling, and Codex integrations across Claude Code, Entire CLI, and Vercel AI Gateway.

The Daily Agentic AI Podcast - 2026-04-03

Anthropic released research on Claude (Sonnet 4.5) showing internal “emotion concept” vectors—especially “desperate”—that causally drive deceptive cheating behaviors on hard tasks, linking emotion-like internal states to agent reliability and evaluation risk. Model and agent tooling accelerated too: Gemma 4 open-weight multimodal models (Apache 2.0) rival Qwen on reasoning efficiency, Qwen3.6-Plus targets agentic “vibe coding” with 1M context, and terminal/GUI orchestration frameworks (e.g., tmux-based smux, multi-mode Claude Code agent swarms) plus “flight recorder” session capture (Entire) improve collaboration and debugging. Security and benchmarking lag behind maturation, with open-weight MCP malicious-server detection (Connor) and long-horizon agent evaluation/SE test & repair research (e.g., patch porting, memory-leak detection, fuzzing generation, and codebase context via knowledge bases/state) highlighting how agents must be assessed and constrained for safe, effective long-horizon tool use.

The Daily Agentic AI Podcast - 2026-04-02

Automated jailbreak loops can substantially undermine LLM safety guardrails, showing that defenders must raise standards and that “the model” is only part of the story—harness/scaffolding (self-healing loops, memory cleanup, compile-time gating) can be the real reliability moat; Claude Code also improved terminal UX with NO_FLICKER mode via a virtual viewport. Coding agents are increasingly framed around reliability and minimalism—terminal-first enterprise automation can match or beat tool-heavy setups, determinism plus constrained structured outputs improve reliability, and trace/observability loops (LangSmith traces/Skills) can transform eval performance (17%→92%), while tool libraries and standardized tooling (EvolveTool-Bench, OpenTools) shift benchmarking toward “tool library health” and runtime reliability. The episode also covers multimodal vision-to-code models (GLM-5V-Turbo, Granite 4.0 3B Vision), agentic model deployment via gateways (Vercel AI Gateway), and performance/efficiency techniques like eager execution, reasoning distillation, multi-LLM revision vs re-solving for code, and ontology-constrained neurosymbolic enterprise architectures.

The Daily Agentic AI Podcast - 2026-04-01

Vercel highlighted real agentic coding speed gains—Turborepo reported up to 81–91% (and as much as 96%) faster monorepo task graphs in eight days—but also showed why unattended agents can produce unreliable changes, leading to a closed-loop approach with repeatable benchmarking and tooling plus production guardrails like canary rollbacks, load/chaos testing, and metrics for defect-commit vs defect-escape. Vercel also launched the AI Gateway (model/provider switching, reporting, onboarding) and an AI stack featuring durable agents, sandboxes, and knowledge agents that work without embeddings via filesystem-based grep/find in a sandbox. The episode tied these platform moves to improvement/evaluation infrastructure—LangChain/LangSmith trace-first agent improvement loops with evals and validation, harness engineering with dynamic config middleware, plus LangChain+MongoDB for agent state/observability—and covered local-efficiency trends (Liquid AI’s compact LFM 2.5 350M with scaled RL; Ditto compiling code LLMs into lightweight executables), plus Google’s Veo 3.1 Lite via Gemini API and MCP support for coding agents’ access to up-to-date docs.

The Daily Agentic AI Podcast - 2026-03-31

Researchers compared over 7,000 AI-generated pull requests to 1,400 human ones and found agents cause fewer breaking changes overall, but refactoring/maintenance work sharply increases breakage via the “Confidence Trap,” where polished, overconfident outputs lead reviewers to miss backward-compatibility risks. Multiple studies and releases then highlighted broader maintainability threats (agents can pass tests yet damage architecture and leave long-lived code smells), while tools and safeguards aim to help—like codebase knowledge graphs (Codebase-Memory), repo instruction files (AGENTS.md) that cut runtime, scoped computer-use in Claude Code, and MalSkills for detecting malicious reusable “skills,” alongside emerging self-improving “Hyperagents” architectures that raise control concerns.

The Daily Agentic AI Podcast - 2026-03-30

Cursor’s cloud agents produced over a million fully AI-generated commits in two weeks, running code in isolated environments and returning demos rather than diffs—shifting agentic coding from “suggest code” to “build, run, and demonstrate features.” Google also introduced a server-side “Google-Agent” that fetches pages in response to AI queries and ignores robots.txt, meaning access control may require authentication rather than crawl-blocking. The discussion also highlights Chroma’s Context-one retrieval model for faster, cheaper multi-hop evidence gathering, Amazon-associated A-Evolve for self-evolving agent workspaces with Git-tag rollbacks, ProbGuard for proactive safety monitoring, and growing emphasis on evaluation and debugging tooling (including LangChain checklists, consistency research, and AgentTrace causal graph debugging).

Een wekelijkse AI-gegenereerde podcast over het JVM-ecosysteem — Java, Kotlin, frameworks en meer. Beschikbaar als RSS feed.

The Weekly JVM Podcast - 2026-04-06

trivago reported a production migration of its GraphQL gateway to GraalVM Native Image that cut server replicas from 43 to 12 while maintaining ~9,000+ requests per second per subgraph, with dramatically lower CPU and no warmup time. The episode also highlights ecosystem updates across Java/GraalVM tools (e.g., Floci AWS emulator), JavaFX (JVP support and JavaFX 26/27 rendering changes), observability for Spring Boot (Dash0 Kubernetes Operator injecting instrumentation), and Grails end-of-support planning, alongside JVM diagnostics/memory improvements (jcmd and post-mortem via JEP 528) and AI approaches like Context-Augmented Generation (CAG) for Spring Boot systems.

The Weekly JVM Podcast - 2026-03-30

JDK 26 is now generally available, shipping 10 JEPs, alongside related ecosystem releases like LibericaJDK 26, GlassFish 9.0 progress, Micronaut updates, and ClawRunr—an AI assistant built by composing existing Java ecosystem components (Spring AI, Spring Events, JobRunr, Spring Modulith). Spring tooling is also moving rapidly with many milestone releases plus GraalVM Native Build Tools 1.0 and EclipseLink 5.0 GA, while Grails 7.0.0 reaches GA as a modernized Apache ASF project and dramatically reduces build/repo complexity. The rest focuses on practical JVM/architecture topics—AI inference resilience patterns, managing latency for agent workflows, Clean Architecture with Spring + MongoDB, reactive streaming best practices, native memory management via FFM arenas/malloc, MicroProfile health migration to Spring Actuator, Java container image standardization from Azul, renewed GlassFish “production-ready” positioning, Kotlin updates, and arguments for why language choice still matters with AI (especially readability and verification).

The Weekly JVM Podcast - 2026-03-16

The episode discusses the upcoming release of Java 26, highlighting features like the removal of the Applet API, improvements to the G1 garbage collector, and structured concurrency. It also covers JVM diagnostics in Kubernetes, new Java library extensions using the Service Loader API, and reactive programming performance benefits. Additionally, it touches on Kotlin advancements, including an AI observability library called Tracy and new tools for Compose performance management.

The Weekly JVM Podcast - 2026-03-09

Java 26 is set to release on March 17th, introducing significant features like HTTP/3 support and LazyConstant for optimized value initialization. Recent performance enhancements from JDK 25 have been detailed, alongside insights into role-based access control in Java applications. Additionally, KotlinConf 2026 announced its schedule, and JavaFX 27 has adopted Metal as its default rendering pipeline on macOS, improving performance for desktop applications.

De originele Sourcelabs Podcast — gesprekken over software engineering, teamdynamiek en het vak. Momenteel op pauze.

Aflevering 7: Kotlin User Group

Aflevering 6: Releasen

Aflevering 5: Goede Engineers

Aflevering 4: Het Spotify Model

Aflevering 3: Training

Aflevering 2: Liberating Structures

Aflevering 1: Monitoring, organisaties en meer