import json import os from typing import Iterator from concurrent.futures import ThreadPoolExecutor, as_completed def remove_duplicate_outputs(input_file_name: str, output_file_name: str, num_threads: int = 10) -> None: """ Removes duplicate entries in the "output" field from the input file and writes the filtered lines to the output file. :param input_file_name: Name of the input JSON file. :param output_file_name: Name of the output JSON file. :param num_threads: Number of threads to use for parallel processing (defaults to 10). """ script_dir = os.path.dirname(os.path.abspath(__file__)) input_file_path = os.path.join(script_dir, input_file_name) output_file_path = os.path.join(script_dir, output_file_name) with open(input_file_path, 'r', encoding='utf-8') as input_fp: lines = input_fp.readlines() with ThreadPoolExecutor(max_workers=num_threads) as executor: futures = [executor.submit(load_and_filter_lines, lines) for _ in range(num_threads)] filtered_lines = [] for future in as_completed(futures): filtered_lines.extend(future.result()) seen_outputs = set() with open(output_file_path, 'w', encoding='utf-8') as output_fp: for line in filtered_lines: try: data = json.loads(line) output_value = data.get('output', '') if output_value not in seen_outputs: seen_outputs.add(output_value) output_fp.write(line) except json.JSONDecodeError: continue def load_and_filter_lines(lines: list[str]) -> list[str]: """ Loads and filters lines from the input dataset. :param lines: List of lines from the input dataset. :return: List of filtered lines. """ filtered_lines = [] for line in lines: try: data = json.loads(line) output_value = data.get('output', '') if output_value: filtered_lines.append(line) except json.JSONDecodeError: continue return filtered_lines # Here is where you put the name of the input file, and name of the file that will be created. Note the input file needs to be in the same directory as the script unless you edit the code to have a file path instead. input_file_name = 'Input_File.json' output_file_name = 'Output_File.json' remove_duplicate_outputs(input_file_name, output_file_name, num_threads=10)