.

Members-Only

Recent Talks & Demos are for members only

Exclusive feed

You must be an AI Tinkerers active member to view these talks and demos.

February 22, 2025 · Seattle

Self-modifying code

The session demonstrates constructing an AI agent that can edit its own source, spawn duplicates, and communicate, covering sandboxing, safety, and practical implementation details.

Overview
Tech stack
  • AWS
    AWS is the world's most comprehensive and broadly adopted cloud platform, offering over 200 fully featured services from 33 geographic Regions.
    AWS is the global leader in cloud infrastructure, delivering over 200 fully featured services. We operate across 105 Availability Zones within 33 geographic Regions, ensuring high availability and low latency for your applications. Core services like Amazon EC2 (virtual servers), Amazon S3 (scalable object storage), and AWS Lambda (serverless compute) provide the foundational building blocks for any workload. This platform allows customers (from startups to Fortune 500s) to innovate faster, reduce operational costs by moving from CapEx to OpEx, and scale instantly. Security remains paramount: we offer 300+ security, compliance, and governance services, meeting standards like ISO 27001 and SOC 1/2/3. Simply put, AWS provides the secure, flexible, and proven foundation you need to build anything.
  • Anthropic
    Anthropic is a frontier AI safety and research company, developing the Claude family of large language models (LLMs) via its Constitutional AI framework.
    Anthropic is an AI safety and research company, founded in 2021 by former OpenAI executives Dario and Daniela Amodei, and structured as a Public Benefit Corporation (PBC) . The core mission is building reliable, steerable AI systems, with a focus on interpretability and long-term alignment . Its flagship product is the Claude family of LLMs, which are highly capable models designed for complex reasoning and coding tasks . A key technical innovation is Constitutional AI (CAI), a training method that aligns the models with a set of ethical principles to ensure helpful, harmless, and honest outputs . The company has secured significant backing, including up to $4 billion from Amazon and a $2 billion commitment from Google .
  • SWE-agent benchmark
    SWE-agent is the state-of-the-art autonomous system enabling LLMs (like GPT-4o) to fix real-world software bugs using developer tools, with performance measured on the SWE-bench dataset.
    SWE-agent is the leading autonomous software engineering system, developed by Princeton and Stanford, that empowers LLMs (like GPT-4o or Claude Sonnet) to resolve real-world GitHub issues. It provides the model with a full suite of developer tools (bash, file editing, search) to generate a correct code patch within a containerized environment. Performance is benchmarked on the industry-standard SWE-bench dataset, which contains 2,294 real GitHub issues. The system is evaluated on its 'resolved rate' (pass@1), which tests both the foundation model and the agentic harness, establishing a rigorous, reproducible measure of AI capability in complex software development.
  • SWE-bench-sonnet
    Claude Sonnet: Anthropic's high-performance AI model, optimized for agentic software engineering tasks and coding workflows.
    SWE-bench-sonnet represents the Claude Sonnet model line's (e.g., Sonnet 4.5) state-of-the-art performance on the SWE-bench Verified benchmark. This evaluation measures an AI's ability to solve real-world GitHub issues from open-source projects. Sonnet 4.5 achieved a record 77.2% on SWE-bench Verified, demonstrating superior autonomous coding capability: it can sustain complex, multi-step reasoning and execute code for long-horizon tasks, making it a powerful foundation for developer-focused AI agents.
  • Sandboxing
    Sandboxing is a security mechanism that executes untrusted code within a tightly controlled, isolated environment to prevent system compromise.
    Sandboxing is a critical security control: it executes untrusted programs (like suspicious email attachments or web browser processes) within a disposable, isolated environment. This containment, often achieved via hypervisor-based virtualization (e.g., Windows Sandbox) or kernel-level user separation (e.g., Android Application Sandbox), strictly limits resource access (file system, network, memory). The core benefit is risk mitigation: any malware 'detonated' inside the sandbox is contained, protecting the host operating system and network infrastructure, and is typically purged upon session closure, ensuring a pristine state for the next use.

Related projects