Datasets:
License:
import multiprocessing | |
import pathlib | |
import traceback | |
from bs4 import BeautifulSoup | |
import orjson | |
import unidecode | |
from RojaStringRemixer import MultilangWikipediaProcessor | |
mp = MultilangWikipediaProcessor() | |
class WikitonaryProcessor(MultilangWikipediaProcessor): | |
def soup_section(self, input_soup: BeautifulSoup, title:str): | |
# Wiktionary is typically single layered | |
languages = [] | |
lang_buffer = [] | |
sub_cat_buffer = [] | |
# print(input_soup) | |
# for html_element in | |
for sect in input_soup.select(".mw-body-content section"): | |
sect.unwrap() | |
[meta.decompose() for meta in input_soup.select(".mw-body-content meta")] | |
[meta.decompose() for meta in input_soup.select(".mw-body-content link")] | |
[meta.decompose() for meta in input_soup.select(".disambig-see-also")] | |
# print(input_soup.selec("section")) | |
# print([i.name for i in list(input_soup.select_one(".mw-body-content").children)]) | |
for html_element in input_soup.select_one(".mw-body-content").children: | |
if html_element.decomposed or html_element is None: | |
continue | |
if html_element.name == "h1": | |
print("[?] Expecting lowest header element to be h1") | |
continue | |
if html_element.name == "h2": | |
# lang_buffer.append(sub_cat_buffer) | |
if sub_cat_buffer: | |
lang_buffer.append(sub_cat_buffer) | |
sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
else: | |
sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
if lang_buffer: | |
languages.append(lang_buffer) | |
lang_buffer = [] | |
elif html_element.name == "h3": | |
lang_buffer.append(sub_cat_buffer) | |
sub_cat_buffer = [self.md.convert_soup(html_element).strip()] | |
else: | |
if html_element.get_text().strip(): | |
sub_cat_buffer.append(self.md.convert_soup(html_element).strip()) | |
if sub_cat_buffer: | |
lang_buffer.append(sub_cat_buffer) | |
if lang_buffer: | |
languages.append(lang_buffer) | |
# print(languages) | |
final_txt = {} | |
for root_lang in languages: | |
lang_data = [] | |
for sub_pair in root_lang[1:]: | |
concat = " \n".join(sub_pair[1:]) | |
lang_data.append(f"## {sub_pair[0]}\n\n{concat}") | |
lang_final = f"# {title.title()}: {root_lang[0][0]}\n\n" + "\n\n".join(lang_data) | |
txt = unidecode.unidecode(lang_final.replace("“","\"").replace("”","\"").replace("\t"," ")).encode('utf-8', errors='replace').decode('unicode_escape', errors='ignore') | |
final_txt[root_lang[0][0].lower()] = txt | |
# print(final_txt) | |
return final_txt | |
all_selectors = [ | |
"style", # Remove styling | |
"sup.reference", # Seems to still exist across | |
"table.nomobile", # Seems to still exist across | |
"div.sister-wikipedia", | |
".floatright", | |
] | |
def convert_soup(self, input_soup: BeautifulSoup): | |
# Remove Generic notices | |
[i.unwrap() for i in input_soup.select('[data-mw^="interface"]')] | |
[i.decompose() for i in input_soup.select(", ".join(self.all_selectors))] | |
[i.decompose() for i in input_soup.select("[class*=\"NavHead\"]")] | |
title = input_soup.select_one("title").extract() | |
for i in input_soup.select(".mw-collapsible"): | |
hidden = i.select_one("div.hidden-content") | |
if hidden: | |
# Expose collapsed content | |
hidden["class"].remove("hidden-content") | |
# Cleanup meta | |
for i in input_soup.select("[data-mw]"): | |
del i["data-mw"] | |
for i in input_soup.select("[rel]"): | |
del i["rel"] | |
for i in input_soup.select("link[href]"): | |
del i["href"] | |
for i in input_soup.select("link[typeof]"): | |
del i["typeof"] | |
for i in input_soup.select("[id]"): | |
del i["id"] | |
for i in input_soup.select("[about]"): | |
del i["about"] | |
for i in input_soup.select("[lang]"): | |
del i["lang"] | |
ifbs = [i.extract() for i in input_soup.select("table.infobox")] | |
ifbs += [i.extract() for i in input_soup.select("table.sidebar.vcard.hlist")] | |
ifbs += [i.extract() for i in input_soup.select("table.infobox.vcard")] | |
ifbs = [self.process_infobox(ifb) for ifb in ifbs] | |
[fig.decompose ()for fig in input_soup.select('figure[typeof^="mw:File/Thumb"]')] | |
return input_soup, ifbs, None, title | |
def convert(self, wiki_data: bytes): | |
data = orjson.loads(wiki_data.rstrip(b"\n")) | |
try: | |
templates = [ | |
":".join(template["name"].split(":")[1:]) | |
for template in data.get("templates", []) | |
] | |
categories = [ | |
":".join(category["name"].split(":")[1:]) | |
for category in data.get("categories", []) | |
] | |
if not data["article_body"].get("wikitext"): | |
return None | |
soup = BeautifulSoup(data["article_body"]["html"], "lxml") | |
is_stub = self.is_stub(soup) | |
soup, infobox, figures, title = self.convert_soup(soup) | |
sections = self.soup_section(soup, data["name"]) | |
text = list(sections.values())[0][2:] | |
return orjson.dumps( | |
{ | |
"id": data["identifier"], | |
"title": data["name"], | |
"url": data["url"], | |
"text": text, | |
"stub": is_stub, | |
"template": templates, | |
"category": categories, | |
"license": [lic["name"] for lic in data["license"]], | |
"wikitext": data["article_body"].get("wikitext"), | |
"lang": data["in_language"]["identifier"], | |
"abstract": data.get("abstract", ""), | |
"sections": sections, | |
"infobox_html": infobox, | |
"figures_dict": figures, | |
} | |
) | |
except Exception as e: | |
print(f"Exception at Soup Conversion: {e} [{data['name']}]\n{traceback.format_exception(e)[-1]}") | |
return None | |
mp = WikitonaryProcessor() | |
fs = [ | |
pathlib.Path(f"wiktionary/enwiktionary_namespace_0_{i}.ndjson") for i in range(16) | |
] | |
# for line in pathlib.Path("") | |
def main(): | |
with open("en-wiktionary.jsonl","wb") as fout: | |
with multiprocessing.Pool(32) as pool: | |
for file in fs: | |
tasks = [] | |
with open(file,"rb") as f: | |
for line in f: | |
tasks.append(pool.apply_async(mp.convert,(line,))) | |
if len(tasks) % 5000 == 0: | |
for task in tasks: | |
r = task.get() | |
if r is None: | |
continue | |
fout.write(r + b"\n") | |
tasks = [] | |
for task in tasks: | |
r = task.get() | |
if r is None: | |
continue | |
fout.write(r + b"\n") | |
# mp.convert(pathlib.Path("misbriefs.json").read_bytes()) | |
main() |