- Human in the Loop
- Posts
- #21 Edition: OpenAI’s Atlas — The First Real AI Browser (and maybe the next OS)
#21 Edition: OpenAI’s Atlas — The First Real AI Browser (and maybe the next OS)
PLUS: Standford's iconic deep learning course is back + IBM open-sources CUGA

Hey, it’s Andreas.
Welcome back to Human in the Loop — your field guide to the latest in AI agents, emerging workflows, and how to stay ahead of what’s here today and what’s coming next.
This week:
→ Google’s Willow quantum chip achieves a 13,000× speed milestone
→ IBM open-sources CUGA — a generalist enterprise agent framework
→ Stanford’s iconic Deep Learning Course returns — led by Andrew Ng & Kian Katanforoosh
Plus: a deep dive into how browsers are quietly becoming the next critical layer in the AI stack — and what OpenAI’s new Atlas release signals for the future.
Let’s dive in.

Weekly Field Notes
🧰 Industry Updates
New drops: Tools, frameworks & infra for AI agents
🔹 Google’s Willow quantum chip achieves 13,000× speed milestone
→ Willow demonstrates quantum error correction at scale — 13,000× faster than classical compute for certain workloads. This could supercharge AI training and optimization in the years ahead.
🌀 LangChain & LangGraph hit v1.0 for production-ready AI agents
→ First two major versions of their open source frameworks just launched.
🌀 Anthropic releases Claude Code on the Web
→ Claude Code is now available in browsers, not just in IDEs, making it easier for more people to use.
🌀 IBM open-sources CUGA, a generalist enterprise agent framework
→ CUGA (Computer Using Generalist Agent) bridges research and enterprise reality. The project has just been open-sourced.
🌀 Manus 1.5 introduces one-prompt full-stack app builder
→ One prompt can now scaffold, design, and deploy a full-stack app.
Manus is extending it’s end-to-end software creation capabilities.
🌀 Brave warns of AI browser security flaw post-Atlas release
→ Shortly after OpenAI’s Atlas launch, Brave issued a warning about agent-based browser vulnerabilities — from cookie hijacking to permission overreach.
🌀 Meta lays off 600 AI staff for leaner, load-bearing teams
→ Meta trims its “ bloated AI division”, aiming for smaller, faster units that focus on efficiency and infra resilience.
🌀 OpenAI adds Company Knowledge to ChatGPT Enterprise
→ Enterprises can now securely inject proprietary data into ChatGPT for personalized reasoning — no retraining required.
🎓 Learning & Upskilling
Sharpen your edge - top free courses this week
📘 Stanford’s Deep Learning Course (Autumn 2025) returns — by Andrew Ng & Kian Katanforoosh
→ The legendary CS230 course is back and freely available on YouTube. Four updated lectures already out, covering the full deep learning lifecycle — from fundamentals to generative models. Still one of the best free ways to learn AI, now refreshed for 2025.
📘 IBM Tutorial — Set up an end-to-end use case for local AI agent with MCP server and watsonx Orchestrate
→ Hands-on guide on connecting a local MCP server to watsonx Orchestrate Developer Edition using Docker Compose.
🌱 Mind Fuel
Strategic reads, enterprise POVs and research
🔹 Stack AI on enterprise AI adoption
→ 99% of Fortune 500s now use AI — but most still chase impact, not experiments. Stack AI’s new paper maps 65+ real agent use cases proving where ROI actually happens.
🔹 Perplexity shares AI workflow playbook for professionals
→ A concise, tactical guide to building personal AI workflows for research, writing, and strategy. Worth bookmarking if you want to integrate AI into your daily routine.
🔹 Microsoft Copilot gets major Human-Centered AI update
→ Major Copilot Fall release focused on making AI more human-centered — with significant improvements in usability, and overall user experience.
🔹 Anthropic published a new study on personalities of AI models
→ New research stress-tested 300K value trade-off prompts across 12 frontier models — revealing distinct moral and behavioral patterns between Anthropic, OpenAI, Google DeepMind, and xAI.

♾️ Thought Loop
What I've been thinking, building, circling this week
Browsers are quietly becoming the most important layer in the AI stack — especially for AI agents.
This week, OpenAI launched ChatGPT Atlas — its own agentic web browser.
Atlas turns ChatGPT from a chat interface into a web-native agent that can browse, act, and execute directly inside pages. It’s a major step beyond conversation — toward autonomous, multi-step reasoning on the open web.
And it’s not a small shift.
The web browser remains the most universal, most used interface we have.
Most of our data, actions, and decisions still live inside it. Traditional chat interfaces can’t handle real workflows, while AI agents need context that’s persistent, real-time, and embedded.
Over 90 % of digital work still happens inside the browser — it’s where we read, write, buy, research, and collaborate.
But OpenAI isn’t alone. Four players are now shaping what the AI browser era might look like:
Core Approach | ChatGPT sidebar + agent for tasks | Single Copilot box; reasons across tabs | AI drives actions and recall by default | Inline answers with agentic workflows |
Autonomous Action | Highly autonomous with agent mode | Lacks personalized autonomy | Very restricted, manual use | Highly autonomous with navigation |
Architecture | Client–server with deep ChatGPT tie-in | Cloud AI inside Chromium Edge | AI-first UX on Chromium | Chromium + Perplexity cloud broker |
Privacy | Page context + memory | Explicit permissions; admin controls | Privacy-by-default, minimal tracking | Context access; publisher-revenue focus |
Strengths | Good for most AI use cases | Good for simple daily interactions | Good UI and recall system | Strong AI search + reasoning |
Limitations | UI doesn’t feel like a browser | Limited AI capabilities | Minimal and lacks agent depth | Reliability issues, hidden memory storage |
Even though Google Chrome still feels untouchable (and started similar capbilities), momentum is building currently. The next platform shift might not be a new app — it might be the browser itself.
Here’s why the browser could become the next AI operating system (especially for AI agents):
1. The browser is the last true interface layer.
→ It’s already where users work — making it the perfect environment for AI (agents) to act without app-switching friction.
2. Agents need context — and the browser has it.
→ Tabs, history, scroll behavior, and DOM data give AI (agents) everything they need to understand intent and act autonomously.
3. Memory lives best in the browser.
→ It spans your entire digital footprint, letting AI (agents) learn, retain, and assist across time and tools.
4. Browsing is the gateway to execution.
→ The browser isn’t just where people consume — it’s where they execute.
That makes it a powerful environment for AI agents to assist with task execution, directly where work happens.

🔧 Tool Spotlight
A tool I'm testing and watching closely this week
nanochat — NanoChat — Andrej Karpathy’s $100 end-to-end ChatGPT clone
Karpathy just open-sourced NanoChat: a minimalist, full-stack ChatGPT-style model you can train from scratch — tokenizer to UI — on a single 8×H100 node in ~4 hours. The entire pipeline runs for under $100, making it one of the most accessible ways to build your own ChatGPT from the ground up.
Highlights
→ Full pipeline: tokenize → pretrain → finetune → serve
→ One script (speedrun.sh) trains a 1.9B model in ~4h
→ Clean, hackable 8K-line codebase (PyTorch + Rust BPE)
→ Built-in web UI + automatic eval report
→ Designed for learning, tinkering, and LLM education
You can try it here: github.com/karpathy/nanochat

That’s it for today. Thanks for reading.
Enjoy this newsletter? Please forward to a friend.
See you next week and have an epic week ahead,
— Andreas

P.S. I read every reply — if there’s something you want me to cover or share your thoughts on, just let me know!
How did you like today's edition? |
