Automated-AI-Web-Researcher.../llm_config.py
2024-11-22 09:04:59 -08:00

55 lines
1.3 KiB
Python

# llm_config.py
LLM_TYPE = "openai" # Options: 'llama_cpp', 'ollama', 'openai'
# LLM settings for llama_cpp
MODEL_PATH = "/home/james/llama.cpp/models/gemma-2-9b-it-Q6_K.gguf"
LLM_CONFIG_LLAMA_CPP = {
"llm_type": "llama_cpp",
"model_path": MODEL_PATH,
"n_ctx": 20000,
"n_gpu_layers": 0,
"n_threads": 8,
"temperature": 0.7,
"top_p": 0.9,
"top_k": 40,
"repeat_penalty": 1.1,
"max_tokens": 1024,
"stop": ["User:", "\n\n"]
}
# LLM settings for Ollama
LLM_CONFIG_OLLAMA = {
"llm_type": "ollama",
"base_url": "http://localhost:11434",
"model_name": "custom-phi3-32k-Q4_K_M",
"temperature": 0.7,
"top_p": 0.9,
"n_ctx": 55000,
"context_length": 55000,
"stop": ["User:", "\n\n"]
}
# New: LLM settings for OpenAI
LLM_CONFIG_OPENAI = {
"base_url": "https://api.openai.com/v1",
"llm_type": "openai",
"model_name": "gpt-4o-mini",
"api_key": "",
"temperature": 0.7,
"top_p": 0.9,
"max_tokens": 4096,
"stop": ["User:", "\n\n"],
"context_length": 128000 # GPT-4 Turbo context window
}
def get_llm_config():
if LLM_TYPE == "llama_cpp":
return LLM_CONFIG_LLAMA_CPP
elif LLM_TYPE == "ollama":
return LLM_CONFIG_OLLAMA
elif LLM_TYPE == "openai":
return LLM_CONFIG_OPENAI
else:
raise ValueError(f"Invalid LLM_TYPE: {LLM_TYPE}")