[LG] In-Context Learning Strategies Emerge Rationally D Wurgaft, E S Lubana, C F Park, H Tanaka... [Stanford University & Harvard University] (2025) https://t.co/hAlrVyRTZl https://t.co/d9fu9R01Nx
[LG] Routing Mamba: Scaling State Space Models with Mixture-of-Experts Projection Z Zhan, L Ren, S Wang, L Liu... [Microsoft] (2025) https://t.co/PrM5zCR0Wu https://t.co/CdTS9LV2eE
[LG] The 4th Dimension for Scaling Model Size R Zhu, H Zhang, T Shi, C Wang... [University of Illinois at Urbana-Champaign & University of Toronto] (2025) https://t.co/gUD1i40ZgY https://t.co/kDYD5XTCO3
Researchers at Sakana AI have developed Text-to-LoRA (T2L), a hypernetwork designed to generate task-specific LoRA adapters for large language models (LLMs) using only natural language descriptions of tasks. Unlike traditional fine-tuning methods that require separate training for each downstream task, T2L can instantly create new LoRA adapters in a single forward pass by conditioning on a text prompt. The system also compresses many existing LoRAs into itself, enabling efficient adaptation of LLMs on the fly. This approach aims to reduce the complexity and computational expense associated with fine-tuning large models. The work was presented at ICML 2025 and authored by researchers including R. Charakorn, E. Cetin, Y. Tang, and R. T. Lange. Additional research highlights include advancements in LLM capabilities such as programming by backpropagation, where models trained solely on source code can execute programs without explicit input-output examples, and emerging strategies in in-context learning. These developments reflect ongoing efforts to improve the adaptability and efficiency of large-scale AI models.