fix: Coderabbit & Recurse

This commit is contained in:
DESKTOP-RTLN3BA\$punk 2025-06-04 17:38:44 -07:00
parent ef252e821e
commit bc1642488f

View file

@ -7,6 +7,7 @@ from app.retriver.documents_hybrid_search import DocumentHybridSearchRetriever
from app.db import SearchSourceConnector, SearchSourceConnectorType, Chunk, Document from app.db import SearchSourceConnector, SearchSourceConnectorType, Chunk, Document
from tavily import TavilyClient from tavily import TavilyClient
from linkup import LinkupClient from linkup import LinkupClient
from sqlalchemy import func
from app.agents.researcher.configuration import SearchMode from app.agents.researcher.configuration import SearchMode
@ -28,7 +29,7 @@ class ConnectorService:
if self.user_id: if self.user_id:
try: try:
# Count total chunks for documents belonging to this user # Count total chunks for documents belonging to this user
from sqlalchemy import func
result = await self.session.execute( result = await self.session.execute(
select(func.count(Chunk.id)) select(func.count(Chunk.id))
.join(Document) .join(Document)
@ -80,7 +81,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(crawled_urls_chunks): for _i, chunk in enumerate(crawled_urls_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -144,7 +145,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(files_chunks): for _i, chunk in enumerate(files_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -356,7 +357,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(slack_chunks): for _i, chunk in enumerate(slack_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -445,7 +446,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(notion_chunks): for _i, chunk in enumerate(notion_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -643,7 +644,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(youtube_chunks): for _i, chunk in enumerate(youtube_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -652,7 +653,7 @@ class ConnectorService:
video_title = metadata.get('video_title', 'Untitled Video') video_title = metadata.get('video_title', 'Untitled Video')
video_id = metadata.get('video_id', '') video_id = metadata.get('video_id', '')
channel_name = metadata.get('channel_name', '') channel_name = metadata.get('channel_name', '')
published_date = metadata.get('published_date', '') # published_date = metadata.get('published_date', '')
# Create a more descriptive title for YouTube videos # Create a more descriptive title for YouTube videos
title = video_title title = video_title
@ -727,7 +728,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(github_chunks): for _i, chunk in enumerate(github_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})
@ -797,7 +798,7 @@ class ConnectorService:
# Process each chunk and create sources directly without deduplication # Process each chunk and create sources directly without deduplication
sources_list = [] sources_list = []
async with self.counter_lock: async with self.counter_lock:
for i, chunk in enumerate(linear_chunks): for _i, chunk in enumerate(linear_chunks):
# Extract document metadata # Extract document metadata
document = chunk.get('document', {}) document = chunk.get('document', {})
metadata = document.get('metadata', {}) metadata = document.get('metadata', {})