SakanaAI has unveiled “Text-to-LoRA,” a hypernetwork that can generate task-specific Low-Rank Adaptation (LoRA) modules for large language models directly from a plain-language description of the desired task. The work, led by Charakorn and colleagues, will be presented at the International Conference on Machine Learning (ICML) 2025. The researchers say the system compresses multiple existing LoRA adapters into a single network and produces new adapters in one forward pass, eliminating the time-consuming and costly fine-tuning generally required to customize large models. By automating adapter creation, Text-to-LoRA aims to let developers deploy domain-specific versions of transformer models almost instantly. The announcement comes amid a broader push to make AI systems self-modifying. A separate framework called Self-Adapting LLMs (SEAL) trains models to generate their own fine-tuning data and weight updates, while other ICML 2025 papers outline techniques such as the Truthfulness Separator Vector for low-cost hallucination detection and refined scaling laws for more efficient training on noisy data. The wave of research highlights accelerating global investment in large-scale AI. According to an analysis shared this week, China now hosts 124 large models compared with nine in the UK, and 32 systems worldwide exceed 10^25 floating-point operations—underscoring industry demand for methods that reduce the cost and complexity of deploying ever-larger language models.
[LG] Ambient Diffusion Omni: Training Good Models with Bad Data G Daras, A Rodriguez-Munoz, A Klivans, A Torralba... [MIT & The University of Texas at Austin] (2025) https://t.co/Pp8XPh3Pvl https://t.co/iXEu7BIKKU
[CL] Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers Y Huang, H Zhu, T Guo, J Jiao... [UC Berkeley] (2025) https://t.co/qyFjMRjuKs https://t.co/6WdT7hg3mK
[CL] Unsupervised Elicitation of Language Models J Wen, Z Ankner, A Somani... [Anthropic] (2025) https://t.co/5K9Pw1WmaD https://t.co/aKrwXY8Qw2