A roundup of highlights in the world of AI: OpenAI changes the way the Atlas browser works, Reddit sues Perplexity for data scraping, Anthropic launches Claude Code in the browser, and Meta introduces a secure hub for AI agents.
OpenAI launches ChatGPT Atlas browser
OpenAI introduced ChatGPT Atlas, a new web browser that directly integrates ChatGPT and agent mode. AI can understand page content, remember context across sessions, and perform tasks without the user leaving their current page. The browser includes an optional „browser memories“ feature that allows ChatGPT to recall details from previously visited pages to provide more personalized assistance; however, users have control over what information is saved or deleted. Atlas also offers an agent mode preview for paid users, allowing ChatGPT to autonomously conduct web research, fill shopping carts or compile documents directly in the browser. Atlas reflects OpenAI's efforts to move toward agent-based AI systems that can handle routine computing tasks, though the company acknowledges risks including bugs and vulnerabilities to malicious instructions. ChatGPT Atlas is available now on macOS for Free, Plus, Pro, and Go users; versions for Windows, iOS, and Android are coming soon.
DeepSeek pilots text-compressing OCR model
DeepSeek has released DeepSeek-OCR, a visio-lingual model that converts text documents into compact visual representations using significantly fewer tokens than the original text. The model achieves 97% precision when compressing text at a 10:1 ratio and maintains 60% accuracy even at 20:1 compression by rendering text as images and encoding them into visual tokens, which the language models then decode back into text. On the OmniDocBench benchmark, DeepSeek-OCR outperforms competing models using significantly fewer tokens - just 100 tokens per page compared to 256 for GOT-OCR2.0 and less than 800 tokens compared to more than 6,000 for MinerU2.0. This compression technique could enable more efficient processing of long contexts in large language models. The code and model weights are publicly available on GitHub.
Claude Code launches web version with parallel agents
Anthropic has released a web version Claude Code, which allows developers to run multiple coding jobs simultaneously across different GitHub repositories directly from the browser. The service runs on a cloud infrastructure managed by Anthropic, with each job running in in an isolated sandbox environment, that includes network and file restrictions to protect code and credentials. As with the command-line and IDE versions, developers can use the Claude Code web interface for bug fixes, routine tasks, testing, backend changes, pull requests, and documentation. This cloud-based approach suggests a shift towards AI agents that handle developer work independently in managed environments. Claude Code for Web is now available in preview for Claude Pro and Max subscribers.
Reddit sues Perplexity AI and other companies for data theft
Reddit has filed a lawsuit against Perplexity AI and three other companies - Oxylabs, AWMProxy and SerpApi - alleging that they illegally downloaded millions of user comments for commercial use. The lawsuit, filed in federal court in New York, accuses the companies of circumventing Reddit's anti-scraping measures and extracting content from Google search results when direct access was blocked. Reddit used a novel technique: it created a test post that could only be crawled by Google's search engine, and then, after data from the post appeared on Perplexity within hours, cited it as evidence. The lawsuit highlights growing tensions over how AI companies procure training data. Perplexity and the other defendant companies deny the allegations and have said they will defend themselves in the lawsuit.
Meta and Hugging Face launch hub for AI agent environments
OpenEnv Hub is a new community platform where developers can create, share and discover standardized environments for AI agents. AI agent environments define the tools, APIs, logins, and execution context an agent needs to perform specific tasks in a secure, sandboxed setting that works for training and deployment. The Hub will be launched soon with initial environments that developers can test, and the OpenEnv 0.1 specification has already been released for community feedback. This initiative addresses a key challenge in AI agent development: large language models need access to appropriate tools, but exposing millions of tools directly is neither safe nor practical. Meta is integrating OpenEnv with its TorchForge RL library and working to extend compatibility with open-source projects including verl, TRL and SkyRL.
GigaBrain-0 uses synthetic data to train robots
The researchers presented GigaBrain-0, a visio-linguistic-action model that trains robots using synthetic data generated by world models, The system generates training scenarios by changing the appearance of objects, their locations, lighting conditions, and camera angles to obtain more diverse training data than most robots get from real-world observation.„embodied Chain-of-Thought“ supervision to divide complex tasks into intermediate steps . Tests on manipulation, long tasks and mobile manipulation showed that GigaBrain-0 outperformed the benchmark model π0 by 10-30 percent . The team also released GigaBrain-0-Small, a lightweight version that runs 10x faster on edge devices while maintaining comparable performance.
The Batch - DeepLearning.AI by Andrew Ng / gnews.cz - GH