|
import subprocess |
|
from datasets import load_dataset, Dataset |
|
import json |
|
from tqdm import tqdm |
|
|
|
|
|
ds = load_dataset("../datasets/the-stack-smol", data_dir='data/typescript') |
|
print('loaded') |
|
|
|
def split_ts_into_chunks(ts_code): |
|
result = subprocess.run( |
|
['node', 'parse_ts.js'], |
|
input=ts_code, |
|
text=True, |
|
) |
|
|
|
if result.returncode != 0: |
|
raise Exception('Error in TypeScript parsing') |
|
|
|
with open('semantic_chunks.jsonl', 'r') as file: |
|
lines = file.read().splitlines() |
|
|
|
chunks = [json.loads(line) for line in lines] |
|
with open('semantic_chunks.jsonl', 'w'): |
|
pass |
|
|
|
return chunks |
|
|
|
|
|
def chunk_ts_file(data): |
|
funcs = split_ts_into_chunks(data['content']) |
|
for i in range(len(funcs)): |
|
funcs[i]['repo'] = data['repository_name'] |
|
funcs[i]['path'] = data['path'] |
|
funcs[i]['language'] = data['lang'] |
|
return funcs |
|
|
|
chunks = [] |
|
for i in tqdm(range(len(ds['train']))): |
|
chunk = chunk_ts_file(ds['train'][i]) |
|
chunks +=(chunk) |
|
if i%100 == 0: |
|
print(len(chunks)) |
|
|
|
dataset = Dataset.from_list(chunks) |
|
print(dataset) |
|
dataset.to_json('ts-chunks.json') |
|
|