
Apple has recently introduced a new series of open-source language models known as OpenELM. These models are designed to be more efficient, requiring 2x fewer pre-training tokens and fewer parameters compared to existing models like OLMo, with which they perform on par. OpenELM also features an open-source training and inference framework, and is capable of layer-wise scaling to efficiently allocate parameters in its transformer model. Additionally, the series includes OpenELM Instruct, which has been described as solid and well-aligned. Apple's advancements in AI are highlighted by their ability to support MLX out of the box and the release of these models as part of their ongoing AI development.







[CL] OpenELM: An Efficient Language Model Family with Open-source Training and Inference Framework S Mehta, M H Sekhavat, Q Cao, M Horton… [Apple] (2024) https://t.co/BtWUAA74wR - OpenELM uses layer-wise scaling to efficiently allocate parameters in the transformer model,… https://t.co/2ab3idHwAc
Apple releases new family of Open-source Efficient Language Models as AI work progresses https://t.co/BgKrzrunNe by @ChanceHMiller
Apple just released 4 open source AI models: Meet OpenELM Instruct - an apple Collection. I am testing now, thus far solid model but very “aligned”. More soon. Article: https://t.co/anfRLAWfKw