All about AI, Web 3.0, BCI
3.26K subscribers
727 photos
26 videos
161 files
3.11K links
This channel about AI, Web 3.0 and brain computer interface(BCI)

owner @Aniaslanyan
Download Telegram
Microsoft unveiled Muse, an AI that can generate minutes of unique game sequences from a single sec of gameplay frames

It's the first World and Human Action Model that predicts 3D environments and actions for playable games.

The scale of training is mind-blowing:

β€” Trained on 1B+ gameplay images
β€” Used 7+ YEARS of continuous gameplay data
β€” Learned from real Xbox multiplayer matches

From a single second of gameplay + controller inputs, Muse can create multiple unique, playable sequences that follow actual game physics, mechanics, and rules.

The version shown in research was trained on just a single game (Bleeding Edge).
Wow, #DeepSeek announced Day 0: Warming up for #OpenSourceWeek

Starting next week, they'll be open-sourcing 5 repos, sharing sincere progress with full transparency.

These humble building blocks in their online service have been documented, deployed and battle-tested in production.

Daily unlocks are coming soon. No ivory towers - just pure garage-energy and community-driven innovation.
πŸ”₯11
Meta presented MLGym: A New Framework and Benchmark for Advancing AI Research Agents

- The first Gym environment for ML tasks
- 13 diverse and open-ended AI research tasks from diverse domains

GitHub
Paper
πŸ‘11
Google dropped SigLIP 2 is the most powerful image-text encoder

SigLIP 2 is new version of SigLIP: best open-source multimodal encoders by Google, now on HF.

What's new?
> Improvements from new masked loss, self-distillation and dense features (better localization)
> Dynamic resolution with Naflex (better OCR).

U can use it to do:
> image-to-image search
> text-to-image-search
> image-to-text search
> image classification with open-ended classes
> train vision language models

SigLIP 2 comes in 3 sizes (base, large, giant), three patch sizes (14, 16, 32) and shape-optimized variants with Naflex.
As usual, supported by transformers from get go.

Models.
πŸ”₯10
Chinese researchers introduced BEAMDOJO

It's a new reinforcement learning framework that teaches robots how to walk on uneven surfaces like stepping stones and balancing beams.

Paper.
❀3
#DeepSeek introduced FlashMLA - efficient MLA decoding kernel for Hopper GPUs, optimized for variable-length sequences and now in production.

- BF16 support
- Paged KV cache (block size 64).
- 3000 GB/s memory-bound & 580 TFLOPS compute-bound on H800
Anthropic to release Claude Sonnet 3.7 on Feb 26

It’s expected to have step-by-step thinking, never before seen coding capabilities and web search.

The best coding model which powers Cursor and Windsurf is about to get a whole lot better.


Claude 3.7 Sonnet is Anthropic's most intelligent model to date and the first Claude model to offer extended thinking - the ability to solve complex problems with careful, step-by-step reasoning.

Anthropic is the first AI lab to introduce a single model where users can balance speed and quality by choosing between standard thinking for near-instant responses or extended thinking or advanced reasoning.

Claude 3.7 Sonnet is state-of-the-art for coding, and delivers advancements in computer use, agentic capabilities, complex reasoning, and content generation. With frontier performance and more control over speed, Claude 3.7 Sonnet is the ideal choice for powering AI agents, especially customer-facing agents, and complex AI workflows.

Supported use cases: RAG or search & retrieval over vast amounts of knowledge, product recommendations, forecasting, targeted marketing, code generation, quality control, parse text from images, agentic computer use, content generation

Model attributes: Reasoning, Text generation, Code generation, Rich text formatting, Agentic computer use
❀‍πŸ”₯7
2 AI agents on a phone call realize they’re both AI and switch to a superior audio signal ggwave

The project mentioned in the demo uses the ggwave library created by developers Anton and Boris.

There's also a related project called "gibberlink" by PennyroyalTea that recently won first place in a hackathon competition.

How It Works?

When the AIs detect they're communicating with another AI rather than a human, they transition to using the ggwave library - a specialized audio signal that allows for much faster and more efficient data transmission through sound waves.

This technology opens up numerous possibilities:

1. Devices can share information through audio channels without requiring internet connectivity.

2. When AI assistants need to communicate with each other, they can do so at vastly improved speeds.

3. Encrypted data can be transmitted through audio in ways less susceptible to conventional interception methods.

4. Speakers, TVs, and other devices can communicate via sound without additional infrastructure.

5. Robots can coordinate activities through audio signals.

6. environments where radio communication is limited or restricted.

7. Systems can exchange supplementary information alongside regular conversation.
πŸ†’6❀5πŸ‘1
#DeepSeek introduced DeepEP - the first open-source EP communication library for MoE model training and inference.

1. Efficient and optimized all-to-all communication
2. Both intranode and internode support with NVLink and RDMA
3. High-throughput kernels for training and inference prefilling
4. Low-latency kernels for inference decoding
5. Native FP8 dispatch support
6. Flexible GPU resource control for computation-communication overlapping.
India has a very vibrant AI ecosystem.
πŸ‘3❀2πŸ”₯2πŸ₯΄1
ROCKET: an AlphaFold augmentation that integrates crystallographic and cryoEM/ET data with room for more

AF-based methods encode rich structural priors but lack a general mechanism for integrating arbitrary data modalities.

ROCKET tackles this by optimizing latent representations to fit experimental data at inference time, without retraining.

Code available soon.
πŸ‘3πŸ”₯3❀2
Google just launched a free version of Gemini Code Assist globally

It comes with:

1. 180K code completions per month

2. Support for all programming languages in the public domain

3. 128K token context window
πŸ‘3πŸ‘3❀2
Nasdaq has filed a 19b-4 filing for the Grayscale POLKADOT ETF.

In 2021, Grayscale announced the establishment of several new cryptocurrency trusts, including the Grayscale Polkadot Trust.
πŸ‘5πŸ‘2πŸ”₯2
AI models now handle voice/speech yet building with them in Python is very frustrating

FastRTC is here to solve
- Automatic Voice Detection
- Handling WebRTC & the backend for real-time apps
- Calling Phones

Github
πŸ”₯5❀2πŸ†’2πŸ‘1
Arc institute introduced the world's largest single-cell dataset

They’re launched the Arc Virtual Cell Atlas, a growing resource for computation-ready single-cell measurements.

As the initial contributions, Vevo Therapeutics has open sourced Tahoe-100M, the world's largest single-cell dataset, mapping 60,000 drug-cell interactions, and announced scBaseCamp, the first RNA sequencing data repository curated using AI agents. Combined, the release includes data from over 300 million cells.
πŸ‘6
#DeepSeek makes 2 major announcements

1. Starting today, DeepSeek is offering significant discounts on their API Platform during off-peak hours (16:30-00:30 UTC daily):
β€’ DeepSeek-V3: 50% OFF
β€’ DeepSeek-R1: Massive 75% OFF

This means you can access powerful AI models at a fraction of the cost during these hours. For example, DeepSeek-R1 output cost drops from $2.19 to just $0.550 per 1M tokens!

2. DeepSeek has also released DeepGEMM - an impressive FP8 GEMM library that supports both dense and MoE GEMMs, powering their V3/R1 models.

Key features:
- Up to 1350+ FP8 TFLOPS on Hopper GPUs
- Lightweight with no heavy dependencies
- Fully Just-In-Time compiled
- Core logic at just ~300 lines of code
- Outperforms expert-tuned kernels on most matrix sizes
- Supports dense layout and two MoE layouts
πŸ”₯8❀6
This is huge: Oklahoma, a state in the USA, has passed the House of Representatives Committee and entered the full vote.

The bill allows the state to invest up to 10% of public funds in BTC or digital assets with a market value of more than $500 billion.
❀3πŸ”₯3πŸ‘2
Good news for devs: Anthropic shipped a more token-efficient tool use implementation for 3.7 Sonnet that uses on average 14% less tokens under-the-hood and shows marked improvement in tool use performance.

Use this beta header: "token-efficient-tools-2025-02-19
❀3πŸ‘3πŸ”₯2