|
import json |
|
import os |
|
from typing import Iterator |
|
from concurrent.futures import ThreadPoolExecutor, as_completed |
|
|
|
def remove_duplicate_outputs(input_file_name: str, output_file_name: str, num_threads: int = 10) -> None: |
|
""" |
|
Removes duplicate entries in the "output" field from the input file and writes the filtered lines to the output file. |
|
|
|
:param input_file_name: Name of the input JSON file. |
|
:param output_file_name: Name of the output JSON file. |
|
:param num_threads: Number of threads to use for parallel processing (defaults to 10). |
|
""" |
|
script_dir = os.path.dirname(os.path.abspath(__file__)) |
|
input_file_path = os.path.join(script_dir, input_file_name) |
|
output_file_path = os.path.join(script_dir, output_file_name) |
|
|
|
with open(input_file_path, 'r', encoding='utf-8') as input_fp: |
|
lines = input_fp.readlines() |
|
|
|
with ThreadPoolExecutor(max_workers=num_threads) as executor: |
|
futures = [executor.submit(load_and_filter_lines, lines) for _ in range(num_threads)] |
|
filtered_lines = [] |
|
for future in as_completed(futures): |
|
filtered_lines.extend(future.result()) |
|
|
|
seen_outputs = set() |
|
with open(output_file_path, 'w', encoding='utf-8') as output_fp: |
|
for line in filtered_lines: |
|
try: |
|
data = json.loads(line) |
|
output_value = data.get('output', '') |
|
if output_value not in seen_outputs: |
|
seen_outputs.add(output_value) |
|
output_fp.write(line) |
|
except json.JSONDecodeError: |
|
continue |
|
|
|
def load_and_filter_lines(lines: list[str]) -> list[str]: |
|
""" |
|
Loads and filters lines from the input dataset. |
|
|
|
:param lines: List of lines from the input dataset. |
|
:return: List of filtered lines. |
|
""" |
|
filtered_lines = [] |
|
for line in lines: |
|
try: |
|
data = json.loads(line) |
|
output_value = data.get('output', '') |
|
if output_value: |
|
filtered_lines.append(line) |
|
except json.JSONDecodeError: |
|
continue |
|
return filtered_lines |
|
|
|
|
|
input_file_name = 'Input_File.json' |
|
output_file_name = 'Output_File.json' |
|
|
|
remove_duplicate_outputs(input_file_name, output_file_name, num_threads=10) |