# Answer Generation This is a generation for 'generation ground truth.' It uses the LLM to generate the answer for the question and the given context (retrieval gt). The answer generation methods can be used in AutoRAG is below. 1. [Basic Generation](#basic-generation) 2. [Concise Generation](#concise-generation) ## Basic Generation This is just a basic generation for the answer. It does not have specific constraints on how it generates the answer. ### OpenAI ```python from autorag.data.beta.schema import QA from autorag.data.beta.generation_gt.openai_gen_gt import make_basic_gen_gt from openai import AsyncOpenAI client = AsyncOpenAI() qa = QA(qa_df) result_qa = qa.batch_apply(make_basic_gen_gt, client=client) ``` ### LlamaIndex ```python from autorag.data.beta.schema import QA from autorag.data.beta.generation_gt.llama_index_gen_gt import make_basic_gen_gt from llama_index.llms.openai import OpenAI llm = OpenAI() qa = QA(qa_df) result_qa = qa.batch_apply(make_basic_gen_gt, llm=llm) ``` ## Concise Generation This is a concise generation for the answer. Concise means that the answer is short and clear, just like a summary. It is usually just a word that is the answer to the question. ### OpenAI ```python from autorag.data.beta.schema import QA from autorag.data.beta.generation_gt.openai_gen_gt import make_concise_gen_gt from openai import AsyncOpenAI client = AsyncOpenAI() qa = QA(qa_df) result_qa = qa.batch_apply(make_concise_gen_gt, client=client) ``` ### LlamaIndex ```python from autorag.data.beta.schema import QA from autorag.data.beta.generation_gt.llama_index_gen_gt import make_concise_gen_gt from llama_index.llms.openai import OpenAI llm = OpenAI() qa = QA(qa_df) result_qa = qa.batch_apply(make_concise_gen_gt, llm=llm) ```