Self Supervised Boy
@selfsupervised
161
subscribers
9
photos
46
links
Reading papers on self/semi/weak supervised DL methods. Papers here:
https://www.notion.so/Self-Supervised-Boy-papers-reading-751aa85ffca948d28feacc45dc3cb0c0
contact me
@martolod
Download Telegram
Join
Self Supervised Boy
161 subscribers
Self Supervised Boy
https://arxiv.org/abs/2509.18542
arXiv.org
Symphony-MoE: Harmonizing Disparate Pre-trained Models into a...
Mixture-of-Experts (MoE) models enable scalable performance by activating large parameter sets sparsely, minimizing computational overhead. To mitigate the prohibitive cost of training MoEs from...
π₯
2
Self Supervised Boy
https://arxiv.org/abs/2509.19170
arXiv.org
Soft Tokens, Hard Truths
The use of continuous instead of discrete tokens during the Chain-of-Thought (CoT) phase of reasoning LLMs has garnered attention recently, based on the intuition that a continuous mixture of...
Self Supervised Boy
https://arxiv.org/abs/2509.21013
arXiv.org
Predicting LLM Reasoning Performance with Small Proxy Model
Given the prohibitive cost of pre-training large language models, it is essential to leverage smaller proxy models to optimize datasets before scaling up. However, this approach becomes...
Self Supervised Boy
https://arxiv.org/abs/2509.26476
arXiv.org
Regression Language Models for Code
We study code-to-metric regression: predicting numeric outcomes of code executions, a challenging task due to the open-ended nature of programming languages. While prior methods have resorted to...
Self Supervised Boy
https://arxiv.org/abs/2510.01123
arXiv.org
Rethinking Thinking Tokens: LLMs as Improvement Operators
Reasoning training incentivizes LLMs to produce long chains of thought (long CoT), which among other things, allows them to explore solution strategies with self-checking. This results in higher...
Self Supervised Boy
https://arxiv.org/abs/2406.18665v4
arXiv.org
RouteLLM: Learning to Route LLMs with Preference Data
Large language models (LLMs) exhibit impressive capabilities across a wide range of tasks, yet the choice of which model to use often involves a trade-off between performance and cost. More...
π
1
Self Supervised Boy
https://arxiv.org/abs/2403.12031
arXiv.org
RouterBench: A Benchmark for Multi-LLM Routing System
As the range of applications for Large Language Models (LLMs) continues to grow, the demand for effective serving solutions becomes increasingly critical. Despite the versatility of LLMs, no...
π
1
Self Supervised Boy
https://arxiv.org/abs/2510.02375
arXiv.org
Pretraining with hierarchical memories: separating long-tail and...
The impressive performance gains of modern language models currently rely on scaling parameters: larger models store more world knowledge and reason better. Yet compressing all world knowledge...
Self Supervised Boy
https://arxiv.org/abs/2510.05445
arXiv.org
AgentRouter: A Knowledge-Graph-Guided LLM Router for Collaborative...
Large language models (LLMs) and agent-based frameworks have advanced rapidly, enabling diverse applications. Yet, with the proliferation of models and agentic strategies, practitioners face...
Self Supervised Boy
https://arxiv.org/abs/2510.12773
arXiv.org
Dr.LLM: Dynamic Layer Routing in LLMs
Large Language Models (LLMs) process every token through all layers of a transformer stack, causing wasted computation on simple queries and insufficient flexibility for harder ones that need...
Self Supervised Boy
https://arxiv.org/abs/2510.18148v1
arXiv.org
Extracting Rule-based Descriptions of Attention Features in Transformers
Mechanistic interpretability strives to explain model behavior in terms of bottom-up primitives. The leading paradigm is to express hidden states as a sparse linear combination of basis vectors,...
π
1
Self Supervised Boy
https://arxiv.org/abs/2510.18147v1
arXiv.org
LLMs Encode How Difficult Problems Are
Large language models exhibit a puzzling inconsistency: they solve complex problems yet frequently fail on seemingly simpler ones. We investigate whether LLMs internally encode problem difficulty...
Self Supervised Boy
https://arxiv.org/abs/2510.21614v1
arXiv.org
Huxley-GΓΆdel Machine: Human-Level Coding Agent Development by an...
Recent studies operationalize self-improvement through coding agents that edit their own codebases. They grow a tree of self-modifications through expansion strategies that favor higher software...