Datasets:

Modalities:
Text
Formats:
json
Languages:
English
Libraries:
Datasets
pandas
License:
json-schema / slurp.py
michaelmior's picture
Add multiple spacing options in schema search
9e31ba2 verified
raw
history blame
2.43 kB
import argparse
import csv
import os
import json
import sys
import time
import requests
import requests_ratelimiter
import tqdm
def slurp(outfile):
query = 'count:all file:\.json$ (content:\'{\n "$schema": "https://json-schema.org/\' or content:\'{\n "$schema": "https://json-schema.org/\')'
session = requests.Session()
adapter = requests_ratelimiter.LimiterAdapter(per_second=2)
session.mount("http://", adapter)
session.mount("https://", adapter)
matches = 0
with session.get(
"https://sourcegraph.com/.api/search/stream",
params={"q": query},
headers={
"Accept": "text/event-stream",
"Authorization": "token " + os.environ["SRC_ACCESS_TOKEN"],
},
stream=True,
) as resp, open(outfile, "w") as f:
pbar = tqdm.tqdm()
writer = csv.writer(f)
writer.writerow(
["repository", "repoStars", "repoLastFetched", "commit", "path"]
)
event = None
for line in resp.iter_lines():
if not line:
continue
time.sleep(0.1)
line = line.decode("utf-8").strip()
if line.startswith("event:"):
event = line.split(":", maxsplit=1)[1].strip()
if event != "matches":
sys.stderr.write(event + "\n")
elif line.startswith("data:"):
data = line.split(":", maxsplit=1)[1].strip()
if event == "filters":
# We don't need to record filtering information
continue
if event == "matches":
record = [
(
m["repository"],
m.get("repoStars", ""),
m.get("repoLastFetched", ""),
m["commit"],
m["path"],
)
for m in json.loads(data)
]
writer.writerows(record)
matches += len(record)
pbar.update(len(record))
elif event == "progress":
sys.stderr.write(data + "\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--outfile", required=True)
args = parser.parse_args()
slurp(args.outfile)