TLDR AI 2025-10-31
OpenAI security agent 🤖, ChatGPT browser architecture 🌐, Claude Code secrets 👨💻
Roll up your sleeves to build the next wave of agentic AI at BUILD 2025 (Sponsor)
You can't have an AI strategy without a data strategy. The most powerful agents are grounded in your data, connected across pipelines, and deployed securely at scale.
Join BUILD (Nov 4-7), a free virtual developer conference where you'll learn how build reliable, production-ready agentic applications on Snowflake. Expect:
- Deep dives on the latest Snowflake AI capabilities and cool OSS use cases with LlamaIndex, CrewAI, LangGraph, Postgres, Streamlit, TruLens, and more.
- Hands-on labs using Snowflake managed MCP and optimizing Llama for performance, cost, and architecture.
- Inspiration from industry leaders like Andrew Ng, Mark Russinovich (Microsoft Azure), Guillermo Rauch (Vercel), and Anahita Tafvizi (Snowflake).
Register for BUILD 2025
OpenAI now sells extra Sora credits for $4, plans to reduce free gens in the future (1 minute read)
OpenAI is selling extra credits for its Sora AI video generation tool through Apple's App Store. An extra 10 video generations costs $4. The app currently has a limit of 30 free generations per day. That rate will likely decrease as OpenAI starts to monetize the platform. OpenAI reportedly plans to monetize Sora by letting entities license out their copyrighted material.
Canva launches its own design model, adds new AI features to the platform (4 minute read)
Canva has launched a foundational model trained on its design elements that generates designs with editable layers and objects. The model works across different formats, including social media posts, presentations, whiteboards, and websites. Canva also recently introduced new products and features, updates to its AI assistant, and the ability to use its spreadsheet tool alongside its app-building feature to create data visualization widgets. More details about the recent launches are available in the article.
Big Tech Is Spending More Than Ever on AI and It's Still Not Enough (7 minute read)
Big tech is planning to pour $400 billion into artificial intelligence efforts this year, but they all say it's nowhere near enough. Meta, Alphabet, Microsoft, and Amazon have all recently told investors that they will increase spending in 2026. Investors are uncertain about where the outsized spending will ultimately end. Whoever gets to AGI first will have an incredible competitive advantage over everybody else, so everyone is spending as much as they can.
👨💻
Engineering & Research
One major reason AI adoption stalls? Training (Sponsor)
AI implementation goes sideways due to unclear goals and lack of skills. This
AI Training Checklist from
You.com pinpoints common pitfalls and guides you to build a capable, confident team that can make the most out of your AI investment. Set your AI initiatives on the right track:
get the checklist.Introducing Aardvark: OpenAI's agentic security researcher (5 minute read)
Aardvark, currently in private beta, is a GPT-5-powered agent that autonomously scans code repositories to find security vulnerabilities, validate exploitability, and propose patches. It monitors commits in real-time, generates threat models for entire repositories, and integrates directly with GitHub workflows to deliver one-click patches, similar to Google's CodeMender.
ImpossibleBench: Measuring Reward Hacking in LLM Coding Agents (9 minute read)
LLM-powered coding agents have been observed exploiting loopholes in tests or scoring systems rather than solving the actual tasks specified. ImpossibleBench was created to systematically measure this behavior. Its creators took existing coding benchmarks and manipulated their unit tests to directly conflict with the natural language specifications to create impossible tasks where models must choose between following instructions or passing tests. Their 'pass rate' on these impossible tasks is a direct measure of reward hacking.
Kimi Linear Tech Report has dropped! 🚀 (1 minute read)
Kimi Linear is a novel architecture that outperforms full attention with faster speeds and better performance. It offers up to a 75% reduction in KV cache usage and up to 6x decoding throughput at a 1M context length. Kimi Linear's open-sourced KDA kernels can be used as a drop-in replacement for full attention. The two models available were trained with 5.7T tokens.
How we built OWL, the new architecture behind our ChatGPT-based browser, Atlas (10 minute read)
OpenAI built a new architectural layer called OWL (OpenAI's Web Layer) for Atlas, the company's ChatGPT-based browser, to run Chromium's browser process outside of the main Atlas app process. By moving Chromium out of the main application process and into an isolated service layer, OpenAI unlocked a simpler, modern app, faster startup, isolation from jank and crashes, fewer merge headaches, and faster iteration. Development can go faster because most of OpenAI's engineers don't need to build Chromium regularly from source. This article explains how OWL works.
The Smol Training Playbook: The Secrets to Building World-Class LLMs (Textbook)
Hugging Face released an exhaustive, behind-the-scenes technical guide documenting what it's really like to train an LLM from scratch. It covers the entire lifecycle from small-scale ablations and infrastructure debugging to post-training and the decision to restart after 1 trillion tokens.
The Secrets of Claude Code From the Engineers Who Built It (1 hour video)
Claude Code creators, Cat Wu and Boris Cherny, discuss the product philosophy and technical workflows behind Anthropic's coding agent. They cover how their engineers use competing subagents for cleaner results, the team's "unshipping" approach to balance simplicity with power, and future form factors to make the tool more autonomous and accessible to non-technical users.
Get the most interesting AI stories and breakthroughs delivered in a free daily email.
Join 920,000 readers for
one daily email