@@ -303,11 +303,26 @@ async def process_query(
303303 context = ingest_query (query )
304304 digest = context .generate_digest ()
305305
306- summary , tree , content = context .generate_digest ()
306+ # Store digest based on S3 configuration
307+ if is_s3_enabled ():
308+ # Upload to S3 instead of storing locally
309+ s3_file_path = generate_s3_file_path (
310+ source = query .url ,
311+ user_name = cast ("str" , query .user_name ),
312+ repo_name = cast ("str" , query .repo_name ),
313+ commit = query .commit ,
314+ include_patterns = query .include_patterns ,
315+ ignore_patterns = query .ignore_patterns ,
316+ )
317+ s3_url = upload_to_s3 (content = context .digest , s3_file_path = s3_file_path , ingest_id = query .id )
318+ # Store S3 URL in query for later use
319+ query .s3_url = s3_url
320+ else :
321+ # Store locally
322+ local_txt_file = Path (clone_config .local_path ).with_suffix (".txt" )
323+ with local_txt_file .open ("w" , encoding = "utf-8" ) as f :
324+ f .write (digest )
307325
308- # Prepare the digest content (tree + content)
309- digest_content = tree + "\n " + content
310- _store_digest_content (query , clone_config , digest_content , summary , tree , content )
311326 except Exception as exc :
312327 _print_error (query .url , exc , max_file_size , pattern_type , pattern )
313328 # Clean up repository even if processing failed
0 commit comments