diff --git a/ai_ta_backend/service/export_service.py b/ai_ta_backend/service/export_service.py index 6eb889c2..726f95d1 100644 --- a/ai_ta_backend/service/export_service.py +++ b/ai_ta_backend/service/export_service.py @@ -35,7 +35,7 @@ def export_documents_json(self, course_name: str, from_date='', to_date=''): response = self.sql.getDocumentsBetweenDates(course_name, from_date, to_date, 'documents') # add a condition to route to direct download or s3 download - if response.count > 1000: + if response.count > 500: # call background task to upload to s3 filename = course_name + '_' + str(uuid.uuid4()) + '_documents.zip' @@ -43,7 +43,8 @@ def export_documents_json(self, course_name: str, from_date='', to_date=''): # background task of downloading data - map it with above ID executor = ProcessPoolExecutor() executor.submit(self.export_data_in_bg, response, "documents", course_name, s3_filepath) - return {"response": 'Download from S3', "s3_path": s3_filepath} + return {"response": 'There are 500+ documents in the export. You will receive an email with the download link shortly!', + "s3_path": s3_filepath} else: # Fetch data @@ -206,13 +207,13 @@ def export_convo_history_json(self, course_name: str, from_date='', to_date=''): from_date (str, optional): The start date for the data export. Defaults to ''. to_date (str, optional): The end date for the data export. Defaults to ''. """ - print("Exporting conversation history to csv file...") + print("Exporting conversation history to json file...") response = self.sql.getDocumentsBetweenDates(course_name, from_date, to_date, 'llm-convo-monitor') - if response.count > 1000: + if response.count > 500: # call background task to upload to s3 - filename = course_name + '_' + str(uuid.uuid4()) + '_convo_history.zip' + filename = course_name + '_' + str(uuid.uuid4()) + '_convo_history.json' s3_filepath = f"courses/{course_name}/{filename}" # background task of downloading data - map it with above ID executor = ProcessPoolExecutor()