Recent discussions among AI experts have highlighted the ongoing debate regarding the use of the lmsys metric by OpenAI. Some analysts argue that lmsys serves as an effective proxy for measuring average user retention and satisfaction with large language models (LLMs), which is crucial for driving revenue and adoption. However, there are concerns about potential overfitting to the lmsys metric, particularly in the context of competing with Google. The term 'overfitting' has been used to describe the risks associated with optimizing performance on the lmsys metric at the expense of broader user experience. Experts are questioning whether this focus could lead to misalignment with actual user needs and preferences, as users may not evaluate LLMs based solely on lmsys performance. The conversation reflects a critical examination of how metrics influence AI development and user engagement.