SurfSense/surfsense_web/hooks/use-llm-configs.ts
2025-06-09 15:50:15 -07:00

246 lines
No EOL
7.1 KiB
TypeScript

"use client"
import { useState, useEffect } from 'react';
import { toast } from 'sonner';
export interface LLMConfig {
id: number;
name: string;
provider: string;
custom_provider?: string;
model_name: string;
api_key: string;
api_base?: string;
litellm_params?: Record<string, any>;
created_at: string;
user_id: string;
}
export interface LLMPreferences {
long_context_llm_id?: number;
fast_llm_id?: number;
strategic_llm_id?: number;
long_context_llm?: LLMConfig;
fast_llm?: LLMConfig;
strategic_llm?: LLMConfig;
}
export interface CreateLLMConfig {
name: string;
provider: string;
custom_provider?: string;
model_name: string;
api_key: string;
api_base?: string;
litellm_params?: Record<string, any>;
}
export interface UpdateLLMConfig {
name?: string;
provider?: string;
custom_provider?: string;
model_name?: string;
api_key?: string;
api_base?: string;
litellm_params?: Record<string, any>;
}
export function useLLMConfigs() {
const [llmConfigs, setLlmConfigs] = useState<LLMConfig[]>([]);
const [loading, setLoading] = useState(true);
const [error, setError] = useState<string | null>(null);
const fetchLLMConfigs = async () => {
try {
setLoading(true);
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/llm-configs/`, {
headers: {
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
method: "GET",
});
if (!response.ok) {
throw new Error("Failed to fetch LLM configurations");
}
const data = await response.json();
setLlmConfigs(data);
setError(null);
} catch (err: any) {
setError(err.message || 'Failed to fetch LLM configurations');
console.error('Error fetching LLM configurations:', err);
} finally {
setLoading(false);
}
};
useEffect(() => {
fetchLLMConfigs();
}, []);
const createLLMConfig = async (config: CreateLLMConfig): Promise<LLMConfig | null> => {
try {
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/llm-configs/`, {
method: 'POST',
headers: {
'Content-Type': 'application/json',
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
body: JSON.stringify(config),
});
if (!response.ok) {
const errorData = await response.json();
throw new Error(errorData.detail || 'Failed to create LLM configuration');
}
const newConfig = await response.json();
setLlmConfigs(prev => [...prev, newConfig]);
toast.success('LLM configuration created successfully');
return newConfig;
} catch (err: any) {
toast.error(err.message || 'Failed to create LLM configuration');
console.error('Error creating LLM configuration:', err);
return null;
}
};
const deleteLLMConfig = async (id: number): Promise<boolean> => {
try {
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/llm-configs/${id}`, {
method: 'DELETE',
headers: {
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
});
if (!response.ok) {
throw new Error('Failed to delete LLM configuration');
}
setLlmConfigs(prev => prev.filter(config => config.id !== id));
toast.success('LLM configuration deleted successfully');
return true;
} catch (err: any) {
toast.error(err.message || 'Failed to delete LLM configuration');
console.error('Error deleting LLM configuration:', err);
return false;
}
};
const updateLLMConfig = async (id: number, config: UpdateLLMConfig): Promise<LLMConfig | null> => {
try {
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/llm-configs/${id}`, {
method: 'PUT',
headers: {
'Content-Type': 'application/json',
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
body: JSON.stringify(config),
});
if (!response.ok) {
const errorData = await response.json();
throw new Error(errorData.detail || 'Failed to update LLM configuration');
}
const updatedConfig = await response.json();
setLlmConfigs(prev => prev.map(c => c.id === id ? updatedConfig : c));
toast.success('LLM configuration updated successfully');
return updatedConfig;
} catch (err: any) {
toast.error(err.message || 'Failed to update LLM configuration');
console.error('Error updating LLM configuration:', err);
return null;
}
};
return {
llmConfigs,
loading,
error,
createLLMConfig,
updateLLMConfig,
deleteLLMConfig,
refreshConfigs: fetchLLMConfigs
};
}
export function useLLMPreferences() {
const [preferences, setPreferences] = useState<LLMPreferences>({});
const [loading, setLoading] = useState(true);
const [error, setError] = useState<string | null>(null);
const fetchPreferences = async () => {
try {
setLoading(true);
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/users/me/llm-preferences`, {
headers: {
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
method: "GET",
});
if (!response.ok) {
throw new Error("Failed to fetch LLM preferences");
}
const data = await response.json();
setPreferences(data);
setError(null);
} catch (err: any) {
setError(err.message || 'Failed to fetch LLM preferences');
console.error('Error fetching LLM preferences:', err);
} finally {
setLoading(false);
}
};
useEffect(() => {
fetchPreferences();
}, []);
const updatePreferences = async (newPreferences: Partial<LLMPreferences>): Promise<boolean> => {
try {
const response = await fetch(`${process.env.NEXT_PUBLIC_FASTAPI_BACKEND_URL}/api/v1/users/me/llm-preferences`, {
method: 'PUT',
headers: {
'Content-Type': 'application/json',
Authorization: `Bearer ${localStorage.getItem('surfsense_bearer_token')}`,
},
body: JSON.stringify(newPreferences),
});
if (!response.ok) {
const errorData = await response.json();
throw new Error(errorData.detail || 'Failed to update LLM preferences');
}
const updatedPreferences = await response.json();
setPreferences(updatedPreferences);
toast.success('LLM preferences updated successfully');
return true;
} catch (err: any) {
toast.error(err.message || 'Failed to update LLM preferences');
console.error('Error updating LLM preferences:', err);
return false;
}
};
const isOnboardingComplete = (): boolean => {
return !!(
preferences.long_context_llm_id &&
preferences.fast_llm_id &&
preferences.strategic_llm_id
);
};
return {
preferences,
loading,
error,
updatePreferences,
refreshPreferences: fetchPreferences,
isOnboardingComplete
};
}