Continual fine-tuning of large language models is critical in dynamic environments where tasks and data distributions change over time. However, this adaptability comes at the cost of catastrophic forgetting, where previously learned skills deteriorate during sequential training. To mitigate this, researchers have proposed Memory-Aware Adaptive Replay (MSSR) for continual LLM fine-tuning1. MSSR aims to balance strong adaptability with retention of previously acquired knowledge by selectively replaying a subset of previously seen data. This approach enables LLMs to rapidly acquire new knowledge while minimizing the degradation of existing skills. The MSSR method has significant implications for the development of more robust and adaptable LLMs, particularly in applications where data distributions shift over time. So what matters to practitioners is that MSSR offers a potential solution to the trade-off between adaptability and knowledge retention, allowing for more effective deployment of LLMs in real-world scenarios.