savvykenya, If you have documents with the answers you're looking for, why not search the documents directly? Why are you embedding the documents then using #RAG (Retrieval Augmenter Generation) to make a large language model give you answers? An LLM generates text, it doesn't search a DB to give you results. So just search the damn DB directly, we already have great search algorithms with O(1) retrieval speeds! #LLMs are so stupid.