OpenAI LLM¶
The openai_llm
module is optimized openai llm module for AutoRAG.
Why use openai_llm
module?¶
There are several advantages using openai
module in AutoRAG.
1. Auto-truncate prompt¶
Sometimes, prompt might exceed a token limitation of the model.
It will occur server-side error, and all your answer results will be gone.
To prevent this, openai_llm
module truncate prompt to the max length of gpt model.
2. Accurate token output¶
In llama_index_llm
module, it does not return proper tokens. It just return pseudo token using GPT2 tokenizer.
When you use openai_llm
module, you can get real tokens that used in gpt model.
In the future, there will be a module that uses token for boosting RAG performance.
3. Accurate log prob output¶
In llama_index_llm
module, it does not return proper log probs since llama index does not support it.
With openai_llm
module, you can get real log probability to every token of generated answers.
In the future, there will be some modules that use log probability, like answer filter.
Module Parameters¶
llm: You can type your ‘model name’ at here. For example,
gpt-4-turbo-2024-04-09
orgpt-3.5-turbo-16k
batch: The batch size of openai api call. You should decrease when you got token limit error.
truncate: Whether you truncate input prompts to model’s max length. Default is True. Recommend you to keep this True.
api_key: OpenAI API key. You can also set this to env variable
OPENAI_API_KEY
.And all parameters from OpenAI Chat Completion without
n
,logprobs
,stream
andtop_logprobs
.
Example config.yaml¶
modules:
- module_type: openai_llm
llm: [ gpt-3.5-turbo, gpt-4-turbo-2024-04-09 ]
temperature: [ 0.1, 1.0 ]
max_tokens: 512