--- myst: html_meta: title: AutoRAG - Multi Query Expansion description: Learn about multi query expansion module in AutoRAG keywords: AutoRAG,RAG,Advanced RAG,query expansion,multi query expansion --- # Multi Query Expansion The `multi_query_expansion` automates the process of prompt tuning by using an LLM to generate multiple queries from different perspectives for a given user input query. The module uses a default multi-query prompt from the [Langchain MultiQueryRetriever](https://python.langchain.com/v0.1/docs/modules/data_connection/retrievers/MultiQueryRetriever/) s default query prompt. ## **Module Parameters** **llm**: The query expansion node requires setting parameters related to our generator modules. - **generator_module_type**: The type of the generator module to use. - **llm**: The type of llm. - Other LLM-related parameters such as `model`, `temperature`, and `max_token` can be set. These are passed as keyword arguments (`kwargs`) to the LLM object, allowing for further customization of the LLM's behavior. **Additional Parameters**: - **prompt**: You can use your own custom prompt for the LLM model. Default prompt comes from langchain MultiQueryRetriever default query prompt. ## **Example config.yaml** ```yaml modules: - module_type: multi_query_expansion generator_module_type: llama_index_llm llm: openai model: [ gpt-3.5-turbo-16k, gpt-3.5-turbo-1106 ] ```