Recent advancements in Large Language Models (LLMs) highlight a growing focus on enhancing their capabilities in processing long contexts and generating synthetic data. A novel approach involves fine-tuning GPT-2 with recurrent memory augmentations, enabling it to handle tasks up to 10M timesteps, addressing the challenge of processing long documents. This development is complemented by the introduction of BABILong, a new benchmark for assessing model performance in long document processing. Additionally, tools like DataDreamer are being utilized for synthetic data generation and reproducible LLM workflows, while LLM Comparator offers visual analytics for side-by-side LLM evaluation. Innovations in subquadratic architectures, such as Linear Transformers with learnable kernel functions, are also being explored to improve in-context learning. Moreover, the Recurrent Memory Transformer (RMT), based on GPT2 with 137M parameters, has shown superior performance over GPT-4 and RAG in contexts up to 128K and even 10M, indicating a significant leap in LLMs' ability to manage extensive context lengths.
LongAgent Scaling Language Models to 128k Context through Multi-Agent Collaboration Large language models (LLMs) have demonstrated impressive performance in understanding language and executing complex reasoning tasks. However, LLMs with long context windows have been notorious… https://t.co/sIc3eYAoeo
A Critical Evaluation of AI Feedback for Aligning Large Language Models Shows that the improvements of the RL step of LLM finetuning are virtually entirely due to the widespread practice of using a weaker teacher model (e.g. GPT-3.5) for SFT data collection than the critic… https://t.co/zaNTkAMnX1
🤿 Deep dive into the best large-language-models out there → https://t.co/5Ltt5eJ9y4 From #GPT4, Bard, RoBERTa + more... here's your definitive list of the top large-language-models pushing the possibilities of #AI and transforming enterprise search. #LLMs #NLProc #GenAI