Abstract: Recent research has explored distilling knowledge from large language models (LLMs) to optimize retriever models, particularly within the retrieval-augmented generation (RAG) framework.