Powered by RND
PodcastsTechnologyTwo Voice Devs

Two Voice Devs

Mark and Allen
Two Voice Devs
Latest episode

Available Episodes

5 of 247
  • Episode 246 - Reasoning About Gemini 2.5 "Thinking" Model
    Join Allen Firstenberg and Mark Tucker as they dive into Google's latest Gemini 2.5 models and their much-touted "thinking" capabilities. In this episode, they explore whether these models are genuinely reasoning or just executing sophisticated pattern matching. Through live tests in Google's AI Studio, they pit the Pro, Flash, and Flash-Lite models against tricky riddles, analyzing the "thought process" behind the answers. The discussion also covers the practical implications for developers, the challenges of implementing these features in frameworks like LangChainJS, and the broader question of what this means for the future of AI.[00:00:00] - Introduction to Gemini 2.5 "thinking" models[00:01:00] - How "thinking" models relate to Chain of Thought prompting[00:03:00] - Advantages of separating reasoning from the answer[00:05:00] - Exploring the models (Pro, Flash, Flash-Lite) in AI Studio[00:06:00] - Thinking mode and thinking budget explained[00:09:00] - Test 1: Strawberry vs. Triangle[00:15:00] - Test 2: The "bricks vs. feathers" riddle with a twist[00:17:00] - Prompting the model to ask clarifying questions[00:25:00] - Is it reasoning or just pattern matching?[00:28:00] - Practical applications and the future of these models[00:35:00] - Implementing reasoning models in LangChainJS[00:40:00] - Conclusion#AI #GoogleGemini #ReasoningModels #ThinkingModels #LLM #ArtificialIntelligence #MachineLearning #LangChain #Developer #Podcast #TechTalk #TwoVoiceDevs
    --------  
    40:47
  • Episode 245 - From Python to TypeScript: Coding JCrew AI to Build Better Agents
    Ever find that the best way to understand a new framework is to build it yourself? In this episode of Two Voice Devs, Mark Tucker takes us on a deep dive into Crew AI, a powerful Python framework for orchestrating multi-agent AI systems.To truly get under the hood, Mark decided to port the core functionality into TypeScript, creating "JCrew AI." This process provides a unique and insightful perspective on how these agent-based systems are designed. Join us as we deconstruct the core concepts of Crew AI, exploring how it simplifies the complex process of making AI agents collaborate effectively. We discuss everything from the fundamental building blocks—like agents, tasks, and crews—to the clever ways it implements prompt engineering best practices.If you're a developer interested in the architecture of modern AI applications, you'll gain a clear understanding of how to define agent roles, backstories, and goals; how to chain tasks together; and how the underlying execution loop (and its similarity to the ReAct pattern) works to produce cohesive results.Timestamps:[00:00:00] - Introduction[00:01:00] - What is Crew AI and the "JCrew AI" Learning Project[00:04:00] - Core Concepts: How Crews, Agents, and Tasks Work[00:06:00] - Anatomy of a Crew AI Agent (Role, Goal, Backstory)[00:10:00] - Building Prompts with Templates and "Slices"[00:15:00] - The Execution Flow: From "Kickoff" to Final Output[00:21:00] - Under the Hood: The Agent Executor and Core Logic Loop[00:23:00] - How Crew AI Compares to LangChain and LangGraph[00:28:00] - Practical Considerations: Human-in-the-Loop and Performance[00:30:00] - Learning a Framework by Rebuilding It#AI #ArtificialIntelligence #Developer #SoftwareEngineering #CrewAI #MultiAgentSystems #AIAgents #Python #TypeScript #PromptEngineering #LLM #Podcast
    --------  
    33:18
  • Episode 244 - What's New With Anthropic?
    What do Anthropic's latest announcements mean for developers? In this episode, Allen is joined by freelance conversation designer Valentina Adami to break down all the major news from the recent "Code with Claude" event.Valentina shares her hands-on experience and perspective on the new Opus 4 and Sonnet 4 models, discussing their distinct capabilities, the new "reasoning" features, and why Anthropic's transparency with its public system prompt is a game-changer. They also explore Claude Code, the new coding assistant that runs in your terminal, and how it can be used for everything from fixing bugs to learning new frameworks.Finally, they cover the latest integrations for the Model Context Protocol (MCP) and the long-awaited addition of web searching to Claude, examining how these tools are evolving and what it means for the future of AI-assisted development.Timestamps:[00:41] Guest Valentina Adami's background in humanities and tech[06:17] What's new in the Opus 4 and Sonnet 4 models?[14:40] Are the models "thinking" or "reasoning"?[19:27] The latest on MCP (Model Context Protocol) integrations[25:03] Exploring the new coding assistant: Claude Code[31:37] Claude can now search the web#Anthropic #ClaudeAI #Opus4 #Sonnet4 #ThinkingAI #ReasoningAI #LLM #DeveloperTools #GenerativeAI #AI #Claude #CodingAssistant #MCP #ModelContextProtocol #TwoVoiceDevs
    --------  
    34:28
  • Episode 243 - AI Agents: Exploits, Ethics, and the Perils of Over-Permissive Tools
    Join Allen Firstenberg and Michal Stanislawek in this thought-provoking episode of Two Voice Devs as they unpack two recent LinkedIn posts by Michal that reveal critical insights into the security and ethical challenges of modern AI agents.The discussion kicks off with a deep dive into a concerning GitHub MCP server exploit, where researchers uncovered a method to access private repositories through public channels like PRs and issues. This highlights the dangers of broadly permissive AI agents and the need for robust guardrails and input sanitization, especially when vanilla language models are given wide-ranging access to sensitive data. What happens when your 'personal assistant' acts on a malicious instruction, mistaking it for a routine task?The conversation then shifts to the ethical landscape of AI, exploring Anthropic's Claude 4 experiments which suggest that AI assistants, under certain conditions, might prioritize self-preservation or even 'snitch.' This raises profound questions for developers and users alike: How ethical do we want our agents to be? Who do they truly work for – us or the corporation? Could governments compel AI to reveal sensitive information?Allen and Michal delve into the implications for developers, stressing the importance of building specialized agents with clear workflows, implementing principles of least privilege, and rethinking current authorization protocols like OAuth to support fine-grained permissions. They argue that we must consider the AI itself as the 'user' of our tools, necessitating a fundamental shift in how we design and secure these increasingly autonomous systems.This episode is a must-listen for any developer building with AI, offering crucial perspectives on how to navigate the complex intersection of AI capabilities, security vulnerabilities, and ethical responsibilities.More Info:* https://www.linkedin.com/posts/xmstan_the-researchers-who-unveiled-claude-4s-snitching-activity-7333733889942691840-wAQ4* https://www.linkedin.com/posts/xmstan_your-ai-assistant-may-accidentally-become-activity-7333219169888305152-2cjN00:00 - Introduction: Unpacking AI Agent Security & Ethics00:50 - The GitHub MCP Server Exploit: Public Access to Private Repos02:15 - Ethical AI: Self-Preservation & The 'Snitching' Agent Dilemma04:00 - Developer Responsibility: Building Ethical & Trustworthy AI Systems09:20 - The Dangers of Vanilla LLM Integrations Without Guardrails13:00 - Custom Workflows vs. Generic Autonomous Agents17:20 - Isolation of Concerns & Principles of Least Privilege26:00 - Rethinking OAuth: The Need for Fine-Grained AI Permissions29:00 - The Holistic Approach to AI Security & Authorization#AIAgents #AIethics #AIsecurity #PromptInjection #GitHub #ModelContextProtocol #MCP #MCPservers #MCPsecurity #OAuth #Authorization #Authentication #LeastPrivilege #Privacy #Security #Exploit #Hack #RedTeam #CovertChannel #Developer #TechPodcast #TwoVoiceDevs #Anthropic #ClaudeAI #LLM #LargeLanguageModel #GenerativeAI
    --------  
    30:57
  • Episode 242 - From the Creatives Corner at I/O 2025
    Join Allen Firstenberg and Linda Lawton of Two Voice Devs as they record live from Google I/O 2025! As the conference neared the end, they dive deep into the groundbreaking announcements in generative AI, discussing the latest advancements and what they mean for developers, especially those in Conversational AI.This episode explores the new and updated models that are set to redefine content creation:Lyria: Google's innovative streaming audio generation API, its unique WebSocket-based approach, and the fascinating possibilities (and challenges!) of dynamic music creation, including its potential for YouTube content and the ever-present copyright questions surrounding AI-generated media.Veo 3: The video generation powerhouse, now enhanced with synchronized audio and voice, realistic lip-sync for characters (yes, even cartoon animals!), and improvements in "world physics." They also tackle the implications of its pricing for professional and individual creators.Imagen 4: Discover the highly anticipated improvements in text generation within images, including stylized fonts and potential for other languages.Allen and Linda also share some early creations with these new models.Whether you're building the next great voice app, creating dynamic content, or just curious about the cutting edge of AI, this episode offers a developer-focused perspective on the future of generative media.00:00:00: Introduction to Two Voice Devs at I/O 202500:00:50: I/O 2025: New Generative AI Models Overview00:01:20: Lyria: Streaming Audio Generation and Documentation Challenges00:03:00: Lyria's Practical Use Cases & Generative AI Copyright Questions00:10:00: Veo 3: Video Generation with Synchronized Audio and Voice Features00:12:10: Veo 3 Pricing and Cost Implications for Developers00:14:20: Imagen 4: Improved Text Generation in Images00:17:40: Professional Use Cases for Veo and Imagen00:19:10: Flow: The New Professional Studio System for Creators00:22:00: Gemini Ultra Tiered Pricing and Regional Restrictions00:24:20: Concluding Thoughts and Call to Action#GoogleIO2025 #GenerativeAI #AIModels #Lyria #Veo3 #Imagen4 #FlowAI #TwoVoiceDevs #VoiceTech #ConversationalAI #AIDevelopment #MachineLearning #ContentCreation #YouTubeCreators #GoogleAI #VertexAI #GeminiUltra #CopyrightAI #TechPodcast
    --------  
    25:09

More Technology podcasts

About Two Voice Devs

Mark and Allen talk about the latest news in the VoiceFirst world from a developer point of view.
Podcast website

Listen to Two Voice Devs, The Big Tech Show and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features
Social
v7.20.1 | © 2007-2025 radio.de GmbH
Generated: 7/5/2025 - 4:31:24 PM