88
99from gitingest .clone import clone_repo
1010from gitingest .ingestion import ingest_query
11- from gitingest .output_formatter import DebugFormatter , DefaultFormatter , SummaryFormatter
11+ from gitingest .output_formatter import DefaultFormatter , SummaryFormatter , generate_digest
1212from gitingest .query_parser import parse_remote_repo
1313from gitingest .utils .git_utils import resolve_commit , validate_github_token
1414from gitingest .utils .logging_config import get_logger
2323 upload_metadata_to_s3 ,
2424 upload_to_s3 ,
2525)
26- from gitingest .schemas import ContextV1
2726from server .server_config import MAX_DISPLAY_SIZE
2827
2928# Initialize logger for this module
@@ -314,17 +313,18 @@ async def process_query(
314313 source = query .url ,
315314 user_name = cast ("str" , query .user_name ),
316315 repo_name = cast ("str" , query .repo_name ),
316+ subpath = query .subpath ,
317317 commit = query .commit ,
318318 include_patterns = query .include_patterns ,
319319 ignore_patterns = query .ignore_patterns ,
320320 )
321- s3_url = upload_to_s3 (content = context . digest , s3_file_path = s3_file_path , ingest_id = query .id )
321+ s3_url = upload_to_s3 (content = generate_digest ( context ) , s3_file_path = s3_file_path , ingest_id = query .id )
322322 # Store S3 URL in query for later use
323323 query .s3_url = s3_url
324324 else :
325325 # Store locally
326326 local_txt_file = Path (clone_config .local_path ).with_suffix (".txt" )
327- print ( f "Writing to { local_txt_file } " )
327+ logger . info ( "Writing digest to local file" , extra = { "file_path" : str ( local_txt_file )} )
328328 with local_txt_file .open ("w" , encoding = "utf-8" ) as f :
329329 f .write (digest )
330330
@@ -340,14 +340,6 @@ async def process_query(
340340 "download full ingest to see more)\n " + digest [:MAX_DISPLAY_SIZE ]
341341 )
342342
343- # _print_success(
344- # url=query.url,
345- # max_file_size=max_file_size,
346- # pattern_type=pattern_type,
347- # pattern=pattern,
348- # summary=digest,
349- # )
350-
351343 digest_url = _generate_digest_url (query )
352344
353345 # Clean up the repository after successful processing
@@ -358,7 +350,7 @@ async def process_query(
358350 short_repo_url = short_repo_url ,
359351 summary = summary ,
360352 digest_url = digest_url ,
361- tree = context .sources [0 ].tree , # TODO: this is a hack to get the tree of the first source
353+ tree = context .sources [0 ].tree , # TODO: this is a hack to get the tree of the first source
362354 content = digest ,
363355 default_max_file_size = max_file_size ,
364356 pattern_type = pattern_type ,
0 commit comments