All about AI, Web 3.0, BCI
3.25K subscribers
727 photos
26 videos
161 files
3.11K links
This channel about AI, Web 3.0 and brain computer interface(BCI)

owner @Aniaslanyan
Download Telegram
Last year Google’s AlphaProof & AlphaGeometry reached a key landmark in AI by achieving silver medal level performance at the International Math Olympiad.

Today, Nature is publishing the methodology behind agent AlphaProof.
🔥6👍2👏2
Anthropic’s applied AI team with a great write up on improving Claude’s frontend design via Skills.

Also with a Claude Code plugin that packages up the skill.
👍4🔥2🥰2
New ByteDance + Yale + NYU + Tsinghua paper builds an LLM based agent called AlphaResearch that searches for new algorithms instead of reusing known ones.

For each problem, AlphaResearch first writes a natural language idea for an algorithm and then turns that idea into code.

The big deal is that this setup lets an LLM push actual mathematical records using a simple loop of scoring ideas and executing code, and the same loop could also search for better algorithms in many other domains.

A reward model trained on peer review data scores each idea and filters out the weakest ones before coding.

An execution engine then runs the code, checks all constraints, and reports a numeric performance score.

The agent loops over this process, sampling old attempts, tweaking ideas and programs, and keeping any version that improves the score.

To measure progress, the authors build a benchmark of 8 open ended algorithm problems with strong human baselines.

On this benchmark, AlphaResearch improves steadily and beats the best human constructions on 2 circle packing tasks, while still trailing people on the other 6.
🔥2🥰2👏2
Czech National Bank has announced the establishment of a pilot digital asset portfolio totaling $1 million, comprising Bitcoin, a USD stablecoin, and a tokenized deposit.

Approved on October 30, the initiative plans to share insights within the next 2–3 years.

The central bank reportedly maintains this is the first instance of a central bank including Bitcoin on its balance sheet.
🔥2🥰2👏2
Google introduced SIMA 2: an agent that plays, reasons, and learns with u in vrtual 3D Worlds

Powered by Gemini, it goes beyond following basic instructions to think, understand, and take actions in interactive environments – meaning you can talk to it through text, voice, or even images.

Google trained SIMA 2 to achieve high-level goals in a wide array of games – allowing it to perform complex reasoning and independently plan how to accomplish tasks.

It acts like a collaborative partner that can explain its intentions and answer questions about its behavior.

SIMA 2 is now far better at carrying out detailed instructions, even in worlds it's never seen before.

It can transfer learned concepts like “mining” in one game and apply it to “harvesting” in another – connecting the dots between similar tasks.

It even navigated unseen environments created in real-time by Genie 3 model.

SIMA 2 can teach itself new skills, learning through trial-and-error, based on feedback from Gemini. Getting better the more it plays –without additional human input.

SIMA 2 research offers a path towards applications in robotics and another step towards AGI in the real world.
3🔥2👏2
OpenAI developed a new way to train small AI models with internal mechanisms that are easier for humans to understand.

Language models like the ones behind ChatGPT have complex, sometimes surprising structures, and we don’t yet fully understand how they work.

In a new research, team train “sparse” models—with fewer, simpler connections between neurons—to see whether their computations become easier to understand.
🔥3🥰3👍2
Efficient Self-Improving Agent Systems. AgentEvolver lets AI agents improve themselves instead of requiring manual prompt tuning.

They use three core mechanisms: self-questioning, self-navigating, and self-attributing.

Agents evaluate their own work, spot failures, and write better instructions for themselves.

This leads to a self-improvement loop capable of running without human oversight.

Shows better performance across benchmarks with less manual work.

The framework works by having agents evaluate their own performance on tasks, identify where they failed or underperformed, and then generate improved behavioral instructions for the next iteration.

The results are impressive.

Agents using this approach show measurable performance gains across diverse benchmarks compared to static configurations, all while reducing the overhead of constant manual optimization.
3🔥3👏2
Google is working on multi-agent systems to help you refine ideas with tournament-like evaluation.

Each run takes around 40 minutes and brings you 100 detailed ideas on a given research topic.

2 new multi-agents are being developed for Gemini Enterprise:
- Idea Generation - "Create a multi-agent innovation session"
- Co-Scientist - "Drive novel scientific discovery with Co-Scientist"

Co-Scientist 3-step workflow 👀
- Tell Co-Scientist what you plan to research, point it to relevant data, and set your evaluation criteria.
- A team of agents will generate ideas on your topic using their available data
- The agents will evaluate the ideas against your criteria and rank them, tournament-style

Google is not only automating research but also preparing a product that will enable others to do so.
7🔥3👏2
Android creator Andy Rubin is launching a new humanoid robotics startup, "Genki Robotics," in Tokyo.

The company is operating in stealth mode, tapping Japan's engineering talent to enter an already crowded field.

During his tenure at Google, Rubin spearheaded an ambitious robotics division, leading the acquisition of numerous startups in 2013, including the high-profile Japanese humanoid firm Shaft, a spin-off from the University of Tokyo.

His interest in legged locomotion, a core challenge in humanoid development, is well-documented. At a 2018 tech conference, Rubin, then leading the incubator Playground Global, predicted a future of "legs everywhere." He argued that legged systems are essential for navigating human-centric environments, such as climbing stairs or using elevators for "last-mile delivery"—tasks impossible for wheeled machines.
🔥4👏4🥰3
MIT and Oxford released their $2,500 agentic AI curriculum on GitHub at no cost.

15,000 people already paid for it.

It covers patterns, orchestration, memory, coordination, and deployment.
A strong roadmap to production ready systems.
5👍4🔥4👏2
Google DeepMind introduced WeatherNext 2 is most advanced system yet, able to generate more accurate and higher-resolution global forecasts.

The model’s improved performance is enabled by a new approach called a Functional Generative Network, which can generate the full range of possible forecasts in a single step.

Team added targeted randomness directly into the architecture, allowing it to explore a wide range of sensible weather scenarios.
🔥5👏43
MIT Introduced JiT (Just image Transformers)

JiTs are simple large-patch Transformers that operate on raw pixels, no tokenizer, pre-training, or extra losses needed.

By predicting clean data on the natural-data manifold, JiT excels in high-dimensional spaces where traditional noise-predicting models can fail.

On ImageNet (256 & 512), JiT achieves competitive generative performance, showing that sometimes going back to basics is the key.

GitHub.
👏43🔥3
Physical intelligence introduced a new model π*0.6

π*0.6 can more than double throughput over a base model trained without RL, and can perform real-world tasks: making espresso drinks, folding diverse laundry, and assembling boxes.

Team trained a general-purpose value function on all of own data, which tells the π*0.6 VLA which actions are good or bad. By asking π*0.6 to produce only good actions, researchers get better performance. Team call this method Recap.

π*0.6 can then collect more autonomous data, which can be used to further train the value function and further improve π*0.6.

During autonomous data collection, a teleoperator can also intervene and provide corrections for significant mistakes, coaching π*0.6 further.

Quantitatively, training π*0.6 with RL can more than double throughput (number of successful task executions per hour) on the hardest tasks and cut the number of failures by as much as a factor of two.
🔥5🥰3👏3
Google DeepMind just released Gemini 3 that helps you learn, build and plan anything.

It comes with state-of-the-art reasoning capabilities, world-leading multimodal understanding, and enables new agentic coding experiences.
🔥7🥰2👏2
PyTorch creator Soumith Chintala has joined Thinking Machines Lab.

Official exit from Meta: Nov 17. New gig at TML: Nov 18.

He says the people there are "incredible" and he is already back to "building new things." The AI talent war continues.
Can LLMs really behave like human investors? How do micro-level behaviors drive macro-level market dynamics?

TwinMarket offers an answer by placing thousands of LLM-driven investors in a realistic stock market environment that incorporates social networks, news, and behavioral biases.

This setup lets us watch bubbles, crashes, and herding emerge from individual decisions.

Calibrated on real market data and grounded in behavioral finance, TwinMarket scales to 1,000+ agents, reproduces key stylized market facts (volatility clustering, fat tails, etc.), and reveals how social interaction and cognitive biases jointly drive systemic risk.

The work is accepted to NeurIPS 2025 and received the Best Paper Award at the ICLR 2025 Financial AI Workshop.

GitHub.
7🔥2👏2
Meta introduced a new generation of Segment Anything Models:

1. SAM 3 enables detecting, segmenting and tracking of objects across images and videos, now with short text phrases and exemplar prompts.

2. SAM 3D brings the model collection into the 3rd dimension to enable precise reconstruction of 3D objects and people from a single 2D image.
🔥2🥰2👏2
Elon Musk’s xAI Introduced Grok 4.1 Fast and the xAI Agent Tools API.

With a 2M context window, it shines in real-world use cases like customer support and deep research.
🆒6
#DeepSeek just released LPLB

Linear-Programming-Based Load Balancer (LPLB) is a parallel load balancer that leverages linear programming to optimize expert parallel workload distribution for MoE (Mixture-of-Experts) models.
🆒4🔥2🥰2👏2
Kimi dropped Seer: Online Context Learning for Fast Synchronous LLM Reinforcement Learning.