All about AI, Web 3.0, BCI
3.25K subscribers
727 photos
26 videos
161 files
3.11K links
This channel about AI, Web 3.0 and brain computer interface(BCI)

owner @Aniaslanyan
Download Telegram
OpenAI released GPT-5-Codex — a version of GPT-5 further optimized for agentic coding in Codex.

Available in the Codex CLI, IDE Extension, web, mobile, and for code reviews in Github.
Anthropic co-founder Jack Clark says in the next 16 months

AI will be smarter than a Nobel prize winner and able to complete tasks that take days, weeks or months.

In short, Jack Clark says, AI will be akin to a “call center of geniuses” or a “country of geniuses.”
😁5🔥4🥰2👏2
ByteDance introduced EMPG, a framework that recalibrates the learning signal using the agent's own uncertainty.

Comparing with GRPO and DAPO, it achieves promising gains on agent benchmarks like WebShop, ALFWorld, & Deep Search.

Paper.
🔥43👏2
Google launched new protocol for agent-driven purchases

Google announced a new open protocol for purchases initiated by AI agents — automated software programs that can shop and make decisions on behalf of users. AI payments protocol supporting credit cards and stablecoins, built with Coinbase, the Ethereum Foundation and over 60 partners, per Fortune. GitHub.

Called the Agent Payments Protocol (AP2), the system is meant to be interoperable between AI platforms, payment systems and vendors, providing a traceable paper trail for each transaction.

In collaboration with cryptocurrency outfits Coinbase, Metamask and the Ethereum foundation, Google also produced an extension that would integrate the cryptocurrency-oriented x402 protocol, allowing for AI-driven purchasing from crypto wallets.
A number of other tech companies are working on their own agentic purchasing systems — most notably Perplexity, which allows for a Buy With Pro service in its agentic browser. The payment provider Stripe also produces software tools for agentic purchasing on its platform, though they are not as comprehensive as AP2.
👍32🥰2👏2
That's a lot of money for robots: Figure has exceeded $1B in funding at a $39B post-money valuation

The round was led by Parkway Venture Capital with significant investments from Brookfield Asset Management, NVIDIA, Macquarie Capital, Intel Capital, Align Ventures, Tamarack Global, LG Technology Ventures, Salesforce, T-Mobile Ventures, and Qualcomm Ventures.

A new funding will support Figure's momentum across three core areas:

1. Scaling humanoid robots into homes & commercial operations

2. Building next-generation GPU infrastructure to accelerate training & simulation

3. Launching advanced data collection efforts for Helix
5🔥4👏2
Tongyi Lab dropped half a dozen new papers, most focused on Deep Research agents.

1. Tongyi DeepResearch: Open-source DeepResearch Agent

• First OSS web agent matching OpenAI’s DeepResearch
• SOTA on HLE (32.9), BrowseComp (43.4/46.7), xbench-DeepSearch (75)
• Full-stack pipeline: Agentic CPT → SFT → RL w/ synthetic data
• Native ReAct & new Heavy Mode (IterResearch) for long-horizon tasks

2. WebResearcher: Unbounded reasoning for long-horizon agents

• IterResearch: Iterative deep-research paradigm (avoids context suffocation & noise)
• WebFrontier: Tool-augmented data engine for complex research tasks
• Parallel agents + synthesis → scalable, evidence-grounded reasoning
• Beats proprietary systems: 36.7% on HLE, 51.7% on BrowseComp

3. AgentScaler: Towards General Agentic Intelligence

• Scales environments for diverse, realistic tool-calling
• Fully simulated envs = verifiable + scalable interactions
• SOTA on τ-bench, τ²-bench, ACEBench
• AgentScaler-30B matches 1T-parameter models with far fewer params

4. AgentFounder: Scaling Agents via Continual Pre-training

• First to propose Agentic CPT → builds agentic foundation models before fine-tuning
• Solves post-training bottlenecks (capabilities + alignment conflict)
• Data synthesis: First-order (planning/actions) + Higher-order (multi-step decision)
• Two-stage training (32K → 128K context)
• SOTA: 39.9% BrowseComp-en, 72.8% GAIA

5. WebWeaver: Structuring Web-Scale Evidence for Deep Research

• Dual-agent framework (Planner + Writer)
• Dynamic outlines: search refine search (human-like loop)
• Memory-grounded, section-by-section synthesis → avoids long-context failures
• SOTA across DeepResearch Bench, DeepConsult, DeepResearchGym
• Produces reliable, well-cited, structured reports

6. ReSum: Long-Horizon Web Agents Without Context Limits

• Problem: ReAct hits context limits in long searches (32k tokens)
• Solution: ReSum periodically compresses history → compact reasoning states
• ReSumTool-30B: specialized summarizer extracts key evidence & gaps
• ReSum-GRPO (RL): trains agents to adapt summaries into reasoning
• +4.5% over ReAct baseline, +8.2% with RL across web search benchmarks.
🔥54👏3
Anthropic shipped two updates for developers using Claude

1. Claude in Xcode 26 Claude Sonnet 4 is now available as a coding assistant directly in Apple's IDE. Developers can connect their Claude account to access natural language code interaction, documentation generation, and inline editing tools. The integration shares usage limits with other Claude platforms and works with Pro, Max, and premium Team/Enterprise plans.

2. Claude Code UX Update A small but useful interface improvement: keywords like "think" and "ultrathink" now get highlighted when they would trigger extended thinking mode. Use /t to disable the mode, preventing accidental activation when these words appear in regular prompts.
🔥32👏2
New a16z benchmark: Which AI-native Office tools actually work?

First, the market splits into two camps:

- Generalists (Assistants: Manus, Genspark; Browsers: Dia, Comet; Extensions: MaxAI, Monica) - flexible but less polished.

- Specialists (Email: Fyxer, Serif; Slides: Gamma, Chronicle; Notes: Mem, Granola) - focused and refined in a single workflow.

a16z benchmarked both across office tasks: summarization, communication, file understanding, research, planning, and execution in 5 use cases.

Key Takeaways from Testing:

1. Specialists still win in PPT, email, and notes - but generalists are catching up, boosted by rapid model progress (e.g. Manus).
2. The horizontal race is heating up - even labs (Anthropic, OpenAI) are entering to own the “work UI”.
3. Convergence is inevitable - verticals are expanding categories, and horizontals are doubling down on use cases.
🔥32👏2
Vitalik Buterin presented Ethereum’s roadmap at Japan Dev Conference today:

1. short-term goals focus on scaling and increasing L1 gas limits;

2. mid-term aims target cross-L2 interoperability and faster responsiveness;

3. long-term vision emphasizes a secure, simple, quantum-resistant, and formally verified minimalist Ethereum.
🆒3👍2
Perceptron AI introduced Isaac 0.1 that can understand and interact with the physical world

Isaac 0.1 is an open-source, 2B params, open weights. Matches or beats models significantly larger on core perception.

Founded by the team behind Meta's Chameleon multimodal models.

Isaac is tuned for where the physical world needs intelligence. Capability alone isn’t enough—you need capability at cost, power, and tail-latency constraints.

In-context learning for perception: show a few annotated examples (defects, safety flags, etc.) and the model adapts in-prompt. No YOLO-style fine-tuning or custom detector stacks.
🔥32👏2
Stanford Introduced Paper2Agent - a system that turns static research papers into interactive AI agents you can chat with and use the tools/data in the paper.

The key idea is to represent a paper and its codebase as a Model Context Protocol (MCP) server, which provides the context for an AI agent to create a paper-specific agent.

GitHub.
🔥32🥰2
MIT Physicists have discovered a new form of magnetism, termed, p-wave magnetism.

This breakthrough paves the way for a new class of ultrafast, compact, energy-efficient, and nonvolatile magnetic memory devices.
4🔥4🥰3
SakanaAI presented Robust Agentic CUDA Kernel Optimization

• Fuses ops, boosts forward/backward passes, outperforms torch baselines

• Agentic LLM pipeline: PyTorch → CUDA → evolutionary runtime optimization

• Soft-verification: LLMs flag incorrect kernels (↑30% verification success)

• robust-kbench: new benchmark for real kernel performance + correctness.

Paper.
👏4🔥3🥰3
Chinese device assembler Luxshare has signed a deal with OpenAI to produce a consumer AI device, possibly the tiny robot.

OpenAI now appears to have multiple separate devices in the works, all planned to arrive next year.
🔥4👍3👏2
DeepSeek introduced DeepSeek-V3.1-Terminus. The latest update builds on V3.1’s strengths while addressing key user feedback.

What’s improved?
1. Language consistency: fewer CN/EN mix-ups & no more random chars.
2. Agent upgrades: stronger Code Agent & Search Agent performance
🔥4👏3🥰2
OpenAI & NVIDIA announced partnership to deploy 10GW of NVIDIA Systems

OpenAI to build & deploy at least 10 gigawatts of AI datacenters with NVIDIA systems representing millions of GPUs for OpenAI’s next-gen AI infrastructure.

To support the partnership, NVIDIA intends to invest up to $100 billion in OpenAI progressively as each gigawatt is deployed.

The first gigawatt of NVIDIA systems will be deployed in the second half of 2026 on NVIDIA’s Vera Rubin platform.
🔥42🥰2
Alibaba introduced Qwen3-Omni — the first natively end-to-end omni-modal AI unifying text, image, audio & video in one model — no modality trade-offs

1. SOTA on 22/36 audio & AV benchmarks
2. 119L text / 19L speech in / 10L speech out
3. 211ms latency | 30-min audio understanding
4. Fully customizable via system prompts
5. Built-in tool calling
6. Open-source Captioner model

What’s Open-Sourced?
- Qwen3-Omni-30B-A3B-Instruct, - Qwen3-Omni-30B-A3B-Thinking, - Qwen3-Omni-30B-A3B Captioner.

GitHub.
HuggingFace.
MS Models.
Demo.
🔥3🥰3😁1
Nvidia presented ReaSyn: Rethinking molecule synthesizability

• Treats synthesis like CoT with Chain-of-Reaction (CoR) steps
• Each reaction = reasoning step → richer supervision & step-by-step learning
• Adds RL finetuning + test-time scaling for better optimization
• SOTA in reconstruction, goal-directed design & hit expansion
• Broad coverage of synthesizable chemical space → practical for drug discovery
Kaggle hosting the 5-Day AI Agents Intensive course with Google on November 10 - 14.

This no-cost course is designed to help you explore the foundations and practical applications of AI agents.
🔥3🥰3👏2