Researchers have introduced MinT, a novel infrastructure system designed to efficiently manage and serve large numbers of low-rank adaptation models, specifically Low-Rank Adaptation (LoRA) post-training. By keeping the base model resident in memory and dynamically loading LoRA adapter revisions, MinT significantly reduces storage requirements and improves serving performance. This approach enables the deployment of millions of specialized models, each tailored to a specific task or policy, without incurring the overhead of storing and managing full checkpoints for each model. The system is particularly suited for applications where a small number of base models are used to generate a large number of derived policies1. This development has significant implications for the scalability and efficiency of large language model (LLM) deployments, allowing for more widespread adoption and application of these models in various domains. This matters to practitioners as it enables the efficient deployment of highly specialized models, which can be critical in applications where model performance and adaptability are key.