All about AI, Web 3.0, BCI – Telegram
All about AI, Web 3.0, BCI
3.22K subscribers
724 photos
26 videos
161 files
3.08K links
This channel about AI, Web 3.0 and brain computer interface(BCI)

owner @Aniaslanyan
Download Telegram
All about AI, Web 3.0, BCI
Future House launched an AI agent Finch that can do bioinformatics analysis, including repeating analysis from research papers. It is multimodal and results in a complete jupyter notebook (python or R) that ends in a concrete conclusion. Starting with closed…
Future House introduced Kosmos, an AI scientist system for data-driven discovery

Kosmos is a multi-agent system designed around a central “world model” to coordinate information across hundreds of scientific agent instances.

Use it.

Given an open-ended objective and dataset, Kosmos can perform up to 12 hours of research to explore, analyze, and complete the objective.

Team presented 7 expert-validated discoveries that Kosmos generated or reproduced across scientific disciplines, including:

1. A novel mechanism of ENT neuron vulnerability with aging
2. Identifying a critical determinant for perovskite performance
3. Evidence that high SOD2 levels may causally reduce myocardial fibrosis.
4🔥4👏4💅1
TSMC broke ground on the world’s most advanced 1.4nm semiconductor fab, a total NT$1.5 trillion (US$48.5 billion) investment in the central Taiwan city of Taichung.

Mass production will start in 2028, with annual revenue seen at NT$500 billion ($16.2 billion).
🔥3👏3🥰2
Can AI invent new math? A new paper from Google DeepMind and renowned mathematician Terence Tao shows how.

Using AlphaEvolve, the team merges LLM-generated ideas with automated evaluation to propose, test, and refine mathematical algorithms.

In tests on 67 problems across analysis, geometry, and number theory, AlphaEvolve not only rediscovered known results but often improved upon them—even generalizing finite cases into universal formulas.

Paired with DeepThink and AlphaProof, it points toward a future where AI doesn’t just assist mathematicians—it collaborates with them in discovery.
🔥82🥰2
Moonshot AI released Kimi K2 Thinking. The Open-Source Thinking Agent Model is here.

- SOTA on HLE (44.9%) and BrowseComp (60.2%)
- Executes up to 200 – 300 sequential tool calls without human interference
- Excels in reasoning, agentic search, and coding
- 256K context window

Built as a thinking agent, K2 Thinking marks our latest efforts in test-time scaling — scaling both thinking tokens and tool-calling turns.

Weights and code.
🔥5😍4🤔2🆒2👏1
Sakana AI is building artificial life and they can evolve: Petri Dish Neural Cellular Automata (PD-NCA) let multiple NCA agents learn and adapt during simulation, not just after training.

Each cell updates its own parameters via gradient descent, turning morphogenesis into a living ecosystem of competing, cooperating, and ever-evolving entities—showing emergent cycles and persistent complexity growth.

GitHub
8🔥3🥰2
DreamGym from Meta is a new framework that lets AI agents train via synthetic reasoning-based experiences instead of costly real rollouts.

It models environment dynamics, replays and adapts tasks, and even improves sim-to-real transfer.

Results: +30% gains on WebArena and PPO-level performance—using only synthetic interactions.
🔥5🥰3👏3
Google Introduced Nested Learning: a new ML paradigm for continual learning that views models as nested optimization problems to enhance long context processing.

A proof-of-concept model, Hope, shows improved performance in language modeling.
🔥63👏3
Alibaba introduced ReasonMed: the largest medical reasoning dataset, advancing LLM performance in clinical QA.

Comprising 370k curated examples distilled from 1.75M reasoning paths, ReasonMed is built through a multi-agent EMD (easy–medium–difficult) pipeline with generation, verification, and an Error Refiner that corrects faulty reasoning steps.

Experiments show that combining detailed CoT reasoning with concise answer summaries yields the most robust fine-tuning outcomes.

- Models trained on ReasonMed redefine the state of the art:
- ReasonMed-7B outperforms all sub-10B models by +4.17% and even beats LLaMA3.1-70B on PubMedQA (+4.60%).
- ReasonMed-14B maintains strong scaling efficiency and competitive accuracy.

Hf.
GitHub.
🆒53👏3🔥1
Moonshot AI : Quantization is not a compromise — it's the next paradigm.

After K2-Thinking's release, many developers have been curious about its native INT4 quantization format.

Moonshot and Zhihu contributor, shared an insider's view on why this choice matters — and why quantization today isn't just about sacrificing precision for speed.

In the context of LLMs, quantization is no longer a trade-off.
With the evolution of param-scaling and test-time-scaling, native low-bit quantization will become a standard paradigm for large model training.

Why Low-bit Quantization Matters?

In modern LLM inference, there are two distinct optimization goals:
• High throughput (cost-oriented): maximize GPU utilization via large batch sizes.
• Low latency (user-oriented): minimize per-query response time.
For Kimi-K2's MoE structure (with 1/48 sparsity), decoding is memory-bound — the smaller the model weights, the faster the compute.
FP8 weights (≈1 TB) already hit the limit of what a single high-speed interconnect GPU node can handle.

By switching to W4A16, latency drops sharply while maintaining quality — a perfect fit for low-latency inference.

Why QAT over PTQ?

Post-training quantization (PTQ) worked well for shorter generations, but failed in longer reasoning chains:
• Error accumulation during long decoding degraded precision.
• Dependence on calibration data caused "expert distortion" in sparse MoE layers.
Thus, K2-Thinking adopted QAT for minimal loss and more stable long-context reasoning.

How it works?

K2-Thinking uses a weight-only QAT with fake quantization + STE (straight-through estimator).
The pipeline was fully integrated in just days — from QAT training → INT4 inference → RL rollout — enabling near lossless results without extra tokens or retraining.

INT4's hidden advantage in RL
Few people mention this: native INT4 doesn't just speed up inference — it accelerates RL training itself.
Because RL rollouts often suffer from "long-tail" inefficiency, INT4's low-latency profile makes those stages much faster.
In practice, each RL iteration runs 10-20% faster end-to-end.
Moreover, quantized RL brings stability: smaller representational space reduces accumulation error, improving learning robustness.

Why INT4, not MXFP4?

Kimi chose INT4 over "fancier" MXFP4/NVFP4 to better support non-Blackwell GPUs, with strong existing kernel support (e.g., Marlin).
At a quant scale of 1×32, INT4 matches FP4 formats in expressiveness while being more hardware-adaptable.
🔥4👏4🥰3
Meta introduced Omnilingual Automatic Speech Recognition (ASR), a suite of models providing ASR capabilities for over 1,600 languages, including 500 low-coverage languages never before served by any ASR system.

While most ASR systems focus on a limited set of languages that are well-represented on the internet, this release marks a major step toward building a truly universal trannoscription system.

They’re released a full suite of models and a dataset:

1. Omnilingual ASR: A suite of ASR models ranging from 300M to 7B parameters, supporting 1600+ languages

2. Omnilingual w2v 2.0: a 7B-parameter multilingual speech representation model that can be leveraged for other downstream speech-related tasks

3. Omnilingual ASR Corpus: a unique dataset spanning 350 underserved languages that was curated in collaboration with our global partners
🔥54👏3
Pleias released a fully synthetic generalist dataset for pretraining, SYNTH and two new SOTA reasoning models exclusively trained on it.

Despite having seen only 200 billion tokens, Baguettotron is currently best-in-class in its size range.

SYNTH is a radical departure from the classic pre-training recipe. At its core it’s an upsampling of Wikipedia 50,000 “vital” articles.

SYNTH is a collection of several synthetic playgrounds: data is not generated through simple prompts but by integrating smaller fine-tuned models into workflows with seeding, constraints, and formal verifications/checks.

Synthetic playgrounds enabled a series of controlled experiments that brought us to favor extreme depth design. Pleias selected a 80-layers architecture for Baguettotron, with improvements across the board on memorization of logical reasoning.

Along with Baguettotron Pleias released the smallest viable language model to date. Monad, a 56M transformer, trained on the English part of SYNTH with non-random performance on MMLU. Desiging Monad an engineering challenge requiring a custom tiny tokenizer.
🔥53🥰3
Google has released a new "Introduction to Agents" guide, which discusses a "self-evolving" agentic system (Level 4).

"At this level, an agentic system can identify gaps in its own capabilities and create new tools or even new agents to fill them."
🔥7🥰2👏2
AELLA is an open-science initiative to make scientific research accessible via structured summaries created by LLMs

Available now:
- Dataset of 100K summaries
- 2 fine-tuned LLMs
- 3d visualizer.

This project spans many disciplines:
- bespoke model-training pipelines
- high-throughput inference systems
- protocols to ensure compute integrity and more.

Models.
Visualizer.
5🔥2👏2
ByteDance launched Doubao-Seed-Code, a model specifically designed for programming tasks.

It supports native 256K long context and has claimed the top spot on the SWE-Bench Verified leaderboard.
A new paper from YANN LECUN. LeJEPA: Provable and Scalable Self-Supervised Learning Without the Heuristics. GitHub.

This could be one of LeCun's last papers at Meta (lol), but it's a really interesting one.

Quick summary:

Yann LeCun's big idea is JEPA, a self-supervised learning method. However, there are various failure modes of this approach, so training strong JEPA models is very brittle, unstable, and quite difficult. So overall JEPA has seen little adoption in practice.

This paper tries to directly address this, making specific design decisions that improve training stability.

The authors identify the isotropic Gaussian as the optimal distribution that JEPA models’ embeddings should follow and design the Sketched Isotropic Gaussian Regularization (SICReg) to constrain embeddings to reach that ideal distribution. This forms the LeJEPA framework, which can be implemented in ~50 lines of code.

On empirical tests, the authors demonstrate stability of training across hyperparameters, architectures, and datasets.

A result particularly interesting to me however is that training a LeJEPA model from scratch directly on the downstream dataset outperforms finetuning a DINOv2/v3 model on the dataset!
Last year Google’s AlphaProof & AlphaGeometry reached a key landmark in AI by achieving silver medal level performance at the International Math Olympiad.

Today, Nature is publishing the methodology behind agent AlphaProof.
🔥6👍2👏2
Anthropic’s applied AI team with a great write up on improving Claude’s frontend design via Skills.

Also with a Claude Code plugin that packages up the skill.
👍4🔥2🥰2