Spaces:
Running
Running
import gradio as gr | |
import base64 | |
import random | |
import gradio as gr | |
#import urllib.request | |
import requests | |
import bs4 | |
import lxml | |
import os | |
#import subprocess | |
from huggingface_hub import InferenceClient,HfApi | |
import random | |
import json | |
import datetime | |
from pypdf import PdfReader | |
import uuid | |
#from query import tasks | |
from gradio_client import Client | |
from agent import ( | |
PREFIX, | |
GET_CHART, | |
COMPRESS_DATA_PROMPT, | |
COMPRESS_DATA_PROMPT_SMALL, | |
LOG_PROMPT, | |
LOG_RESPONSE, | |
) | |
api=HfApi() | |
client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1") | |
def sort_fn(inp): | |
client_sort = Client("Omnibus/sort_document") | |
sen,nouns = client_sort.predict( | |
f"{inp}", # str in 'Paste Text' Textbox component | |
api_name="/sort_doc" | |
) | |
return nouns | |
def find_all(url): | |
return_list=[] | |
print (url) | |
#if action_input in query.tasks: | |
print (f"trying URL:: {url}") | |
try: | |
if url != "" and url != None: | |
out = [] | |
source = requests.get(url) | |
#source = urllib.request.urlopen(url).read() | |
soup = bs4.BeautifulSoup(source.content,'lxml') | |
rawp=(f'RAW TEXT RETURNED: {soup.text}') | |
cnt=0 | |
cnt+=len(rawp) | |
out.append(rawp) | |
out.append("HTML fragments: ") | |
q=("a","p","span","content","article") | |
for p in soup.find_all("a"): | |
out.append([{"LINK TITLE":p.get('title'),"URL":p.get('href'),"STRING":p.string}]) | |
print(rawp) | |
return True, rawp | |
else: | |
return False, "Enter Valid URL" | |
except Exception as e: | |
print (e) | |
return False, f'Error: {e}' | |
#else: | |
# history = "observation: The search query I used did not return a valid response" | |
return "MAIN", None, history, task | |
FIND_KEYWORDS="""Find keywords from the dictionary of provided keywords that are relevant to the users query. | |
Return the keyword:value pairs from the list in the form of a JSON file output. | |
dictionary: | |
{keywords} | |
user query: | |
""" | |
def find_keyword_fn(c,inp,data): | |
data=str(data) | |
seed=random.randint(1,1000000000) | |
divr=int(c)/20000 | |
divi=int(divr)+1 if divr != int(divr) else int(divr) | |
chunk = int(int(c)/divr) | |
out = [] | |
s=0 | |
e=chunk | |
print(f'e:: {e}') | |
#task = f'Compile this data to fulfill the task: {task}, and complete the purpose: {purpose}\n' | |
for z in range(divi): | |
print(f's:e :: {s}:{e}') | |
hist = data[s:e] | |
resp = run_gpt( | |
FIND_KEYWORDS, | |
stop_tokens=[], | |
max_tokens=8192, | |
seed=seed, | |
keywords=data, | |
).strip("\n") | |
out.append(resp) | |
#new_history = resp | |
print (resp) | |
#out+=resp | |
e=e+chunk | |
s=s+chunk | |
return out | |
def read_txt(txt_path): | |
text="" | |
with open(txt_path,"r") as f: | |
text = f.read() | |
f.close() | |
print (text) | |
return text | |
def read_pdf(pdf_path): | |
text="" | |
reader = PdfReader(f'{pdf_path}') | |
number_of_pages = len(reader.pages) | |
for i in range(number_of_pages): | |
page = reader.pages[i] | |
text = f'{text}\n{page.extract_text()}' | |
print (text) | |
return text | |
error_box=[] | |
def read_pdf_online(url): | |
uid=uuid.uuid4() | |
print(f"reading {url}") | |
response = requests.get(url, stream=True) | |
print(response.status_code) | |
text="" | |
################# | |
##################### | |
try: | |
if response.status_code == 200: | |
with open("test.pdf", "wb") as f: | |
f.write(response.content) | |
#f.close() | |
#out = Path("./data.pdf") | |
#print (out) | |
reader = PdfReader("test.pdf") | |
number_of_pages = len(reader.pages) | |
print(number_of_pages) | |
for i in range(number_of_pages): | |
page = reader.pages[i] | |
text = f'{text}\n{page.extract_text()}' | |
print(f"PDF_TEXT:: {text}") | |
return text | |
else: | |
text = response.status_code | |
error_box.append(url) | |
print(text) | |
return text | |
except Exception as e: | |
print (e) | |
return e | |
VERBOSE = True | |
MAX_HISTORY = 100 | |
MAX_DATA = 20000 | |
def format_prompt(message, history): | |
prompt = "<s>" | |
for user_prompt, bot_response in history: | |
prompt += f"[INST] {user_prompt} [/INST]" | |
prompt += f" {bot_response}</s> " | |
prompt += f"[INST] {message} [/INST]" | |
return prompt | |
def run_gpt_no_prefix( | |
prompt_template, | |
stop_tokens, | |
max_tokens, | |
seed, | |
**prompt_kwargs, | |
): | |
print(seed) | |
try: | |
generate_kwargs = dict( | |
temperature=0.9, | |
max_new_tokens=max_tokens, | |
top_p=0.95, | |
repetition_penalty=1.0, | |
do_sample=True, | |
seed=seed, | |
) | |
content = prompt_template.format(**prompt_kwargs) | |
#if VERBOSE: | |
print(LOG_PROMPT.format(content)) | |
#formatted_prompt = format_prompt(f"{system_prompt}, {prompt}", history) | |
#formatted_prompt = format_prompt(f'{content}', history) | |
stream = client.text_generation(content, **generate_kwargs, stream=True, details=True, return_full_text=False) | |
resp = "" | |
for response in stream: | |
resp += response.token.text | |
#yield resp | |
#if VERBOSE: | |
print(LOG_RESPONSE.format(resp)) | |
return resp | |
except Exception as e: | |
print(f'no_prefix_error:: {e}') | |
return "Error" | |
def run_gpt( | |
prompt_template, | |
stop_tokens, | |
max_tokens, | |
seed, | |
**prompt_kwargs, | |
): | |
print(seed) | |
timestamp=datetime.datetime.now() | |
generate_kwargs = dict( | |
temperature=0.9, | |
max_new_tokens=max_tokens, | |
top_p=0.95, | |
repetition_penalty=1.0, | |
do_sample=True, | |
seed=seed, | |
) | |
content = PREFIX.format( | |
timestamp=timestamp, | |
purpose="Compile the provided data and complete the users task" | |
) + prompt_template.format(**prompt_kwargs) | |
#if VERBOSE: | |
print(LOG_PROMPT.format(content)) | |
#formatted_prompt = format_prompt(f"{system_prompt}, {prompt}", history) | |
#formatted_prompt = format_prompt(f'{content}', history) | |
stream = client.text_generation(content, **generate_kwargs, stream=True, details=True, return_full_text=False) | |
resp = "" | |
for response in stream: | |
resp += response.token.text | |
#yield resp | |
if VERBOSE: | |
print(LOG_RESPONSE.format(resp)) | |
return resp | |
def compress_data(c, instruct, history): | |
seed=random.randint(1,1000000000) | |
print (c) | |
#tot=len(purpose) | |
#print(tot) | |
divr=int(c)/MAX_DATA | |
divi=int(divr)+1 if divr != int(divr) else int(divr) | |
chunk = int(int(c)/divr) | |
print(f'chunk:: {chunk}') | |
print(f'divr:: {divr}') | |
print (f'divi:: {divi}') | |
out = [] | |
#out="" | |
s=0 | |
e=chunk | |
print(f'e:: {e}') | |
new_history="" | |
#task = f'Compile this data to fulfill the task: {task}, and complete the purpose: {purpose}\n' | |
for z in range(divi): | |
print(f's:e :: {s}:{e}') | |
hist = history[s:e] | |
resp = run_gpt( | |
COMPRESS_DATA_PROMPT_SMALL, | |
stop_tokens=["observation:", "task:", "action:", "thought:"], | |
max_tokens=8192, | |
seed=seed, | |
direction=instruct, | |
knowledge="", | |
history=hist, | |
).strip("\n") | |
out.append(resp) | |
#new_history = resp | |
print (resp) | |
#out+=resp | |
e=e+chunk | |
s=s+chunk | |
return out | |
def compress_data_og(c, instruct, history): | |
seed=random.randint(1,1000000000) | |
print (c) | |
#tot=len(purpose) | |
#print(tot) | |
divr=int(c)/MAX_DATA | |
divi=int(divr)+1 if divr != int(divr) else int(divr) | |
chunk = int(int(c)/divr) | |
print(f'chunk:: {chunk}') | |
print(f'divr:: {divr}') | |
print (f'divi:: {divi}') | |
out = [] | |
#out="" | |
s=0 | |
e=chunk | |
print(f'e:: {e}') | |
new_history="" | |
#task = f'Compile this data to fulfill the task: {task}, and complete the purpose: {purpose}\n' | |
for z in range(divi): | |
print(f's:e :: {s}:{e}') | |
hist = history[s:e] | |
resp = run_gpt( | |
COMPRESS_DATA_PROMPT, | |
stop_tokens=["observation:", "task:", "action:", "thought:"], | |
max_tokens=8192, | |
seed=seed, | |
direction=instruct, | |
knowledge=new_history, | |
history=hist, | |
).strip("\n") | |
new_history = resp | |
print (resp) | |
out+=resp | |
e=e+chunk | |
s=s+chunk | |
''' | |
resp = run_gpt( | |
COMPRESS_DATA_PROMPT, | |
stop_tokens=["observation:", "task:", "action:", "thought:"], | |
max_tokens=8192, | |
seed=seed, | |
direction=instruct, | |
knowledge=new_history, | |
history="All data has been recieved.", | |
)''' | |
print ("final" + resp) | |
#history = "observation: {}\n".format(resp) | |
return resp | |
def get_chart(inp): | |
seed=random.randint(1,1000000000) | |
try: | |
resp = run_gpt_no_prefix( | |
GET_CHART, | |
stop_tokens=[], | |
max_tokens=8192, | |
seed=seed, | |
inp=inp, | |
).strip("\n") | |
print(resp) | |
except Exception as e: | |
print(f'Error:: {e}') | |
resp = e | |
return resp | |
def format_json(inp): | |
print("FORMATTING:::") | |
print(type(inp)) | |
print("###########") | |
print(inp) | |
print("###########") | |
print("###########") | |
new_str="" | |
matches=["```","#","//"] | |
for i,line in enumerate(inp): | |
line = line.strip() | |
print(line) | |
#if not any(x in line for x in matches): | |
new_str+=line.strip("\n").strip("```").strip("#").strip("//") | |
print("###########") | |
print("###########") | |
#inp = inp.strip("<\s>") | |
new_str=new_str.strip("</s>") | |
out_json=eval(new_str) | |
print(out_json) | |
print("###########") | |
print("###########") | |
return out_json | |
this=["1.25"] | |
css=""" | |
#wrap { width: 100%; height: 100%; padding: 0; overflow: auto; } | |
#frame { width: 100%; border: 1px solid black; } | |
#frame { zoom: $ZOOM; -moz-transform: scale($ZOOM); -moz-transform-origin: 0 0; } | |
""" | |
def mm(graph,zoom): | |
code_out="" | |
for ea in graph.split("\n"): | |
code=ea.strip().strip("\n") | |
code_out+=code | |
#out_html=f'''<div><iframe src="https://omnibus-mermaid-script.static.hf.space/index.html?mermaid={code_out}&rand={random.randint(1,1111111111)}" height="500" width="500"></iframe></div>''' | |
#url=f"https://omnibus-mermaid-script.static.hf.space/index.html?mermaid={code_out}" | |
url=f"https://omnibus-text-to-chart.hf.space/file=merm.html?mermaid={code_out}" | |
out_html=f'''<div id="wrap" style="width: 100%; height: 100%;max-height:600px; padding: 0; overflow: auto;"><iframe id="frame" src="{url}" style="width:100%; height:600px; border: 1px solid black; zoom: {str(zoom)}; -moz-transform: scale({str(zoom)}); -moz-transform-origin: 0 0;" allow="accelerometer; ambient-light-sensor; autoplay; battery; camera; document-domain; encrypted-media; fullscreen; geolocation; gyroscope; layout-animations; legacy-image-formats; magnetometer; microphone; midi; oversized-images; payment; picture-in-picture; publickey-credentials-get; sync-xhr; usb; vr ; wake-lock; xr-spatial-tracking" sandbox="allow-forms allow-modals allow-popups allow-popups-to-escape-sandbox allow-same-origin allow-scripts allow-downloads"></iframe></div>''' | |
return out_html,url | |
def summarize(inp,history,data=None,files=None,directory=None,url=None,pdf_url=None,pdf_batch=None): | |
json_box=[] | |
chart_out="" | |
if inp == "": | |
inp = "Process this data" | |
history.clear() | |
history = [(inp,"Working on it...")] | |
yield "",history,chart_out,chart_out,json_box,"" | |
if pdf_batch.startswith("http"): | |
lab="PDF Batch" | |
c=0 | |
data="" | |
for i in str(pdf_batch): | |
if i==",": | |
c+=1 | |
print (f'c:: {c}') | |
try: | |
for i in range(c+1): | |
batch_url = pdf_batch.split(",",c)[i] | |
bb = read_pdf_online(batch_url) | |
data=f'{data}\nFile Name URL ({batch_url}):\n{bb}' | |
except Exception as e: | |
print(e) | |
#data=f'{data}\nError reading URL ({batch_url})' | |
if directory: | |
lab="Directory" | |
for ea in directory: | |
print(ea) | |
if pdf_url.startswith("http"): | |
lab="PDF URL" | |
print("PDF_URL") | |
out = read_pdf_online(pdf_url) | |
data=out | |
if url.startswith("http"): | |
lab="Raw HTML" | |
val, out = find_all(url) | |
if not val: | |
data="Error" | |
rawp = str(out) | |
else: | |
data=out | |
if files: | |
lab="Files" | |
for i, file in enumerate(files): | |
try: | |
print (file) | |
if file.endswith(".pdf"): | |
zz=read_pdf(file) | |
print (zz) | |
data=f'{data}\nFile Name ({file}):\n{zz}' | |
elif file.endswith(".txt"): | |
zz=read_txt(file) | |
print (zz) | |
data=f'{data}\nFile Name ({file}):\n{zz}' | |
except Exception as e: | |
data=f'{data}\nError opening File Name ({file})' | |
print (e) | |
if data != "Error" and data != "": | |
history.clear() | |
history = [(inp,f"Data: Loaded, processing...")] | |
yield "",history,chart_out,chart_out,json_box,"" | |
print(inp) | |
out = str(data) | |
rl = len(out) | |
print(f'rl:: {rl}') | |
c=1 | |
for i in str(out): | |
if i == " " or i=="," or i=="\n": | |
c +=1 | |
print (f'c:: {c}') | |
json_out = compress_data(c,inp,out) | |
out = str(json_out) | |
try: | |
json_out=format_json(json_out) | |
except Exception as e: | |
print (e) | |
history.clear() | |
history = [(inp,"Building Chart...")] | |
yield "",history,chart_out,chart_out,json_out,"" | |
chart_out = get_chart(str(json_out)) | |
chart_list=chart_out.split("\n") | |
go=True | |
cnti=1 | |
line_out="" | |
for ii, line in enumerate(chart_list): | |
if go: | |
line=line.strip().replace('"',"") | |
if "```" in chart_list[ii]: | |
while True: | |
line_out+=chart_list[ii+cnti].strip().strip("\n").replace('"',"") | |
if not line_out.endswith(";"): | |
line_out+=";" | |
line_out+="\n" | |
cnti+=1 | |
if "```" in chart_list[ii+cnti]: | |
go=False | |
break | |
chart_html,chart_url=mm(line_out,1) | |
#print(chart_out) | |
else: | |
rawp = "Provide a valid data source" | |
history.clear() | |
history.append((inp,chart_out)) | |
yield "", history,chart_html,line_out,json_out,chart_url | |
################################# | |
def clear_fn(): | |
return "",[(None,None)] | |
def create_image(url): | |
#source = requests.get(url) | |
#source = urllib.request.urlopen(url).read() | |
#soup = bs4.BeautifulSoup(source.content,'lxml') | |
#rawp=(f'RAW TEXT RETURNED: {soup.text}') | |
#cnt=0 | |
#cnt+=len(rawp) | |
#out.append(rawp) | |
#out.append("HTML fragments: ") | |
#q=("a","p","span","content","article") | |
#out=[] | |
#for b in soup.find_all("div"): | |
# for p in b.find_all("pre, {'class': 'mermaid'}"): | |
# out.append(p.find('svg')) | |
print(url) | |
#out.append(p.string) | |
with open("tmp.svg","w") as svg: | |
svg.write(url) | |
return "tmp.svg" | |
score_js=""" | |
function(text_input) { | |
console.log(text_input); | |
const iframe = document.getElementById("frame").contentWindow.document.getElementById('chart').innerHTML; | |
console.log(iframe); | |
return [iframe]; | |
} | |
""" | |
def zoom_update(inp): | |
this.clear() | |
this.append(str(inp)) | |
return gr.update() | |
with gr.Blocks() as app: | |
gr.HTML("""<center><h1>Text -to- Chart</h1><h3>Mixtral 8x7B</h3>""") | |
chatbot = gr.Chatbot(label="Mixtral 8x7B Chatbot",show_copy_button=True) | |
with gr.Row(): | |
with gr.Column(scale=3): | |
prompt=gr.Textbox(label = "Instructions (optional)") | |
with gr.Column(scale=1): | |
button=gr.Button() | |
#models_dd=gr.Dropdown(choices=[m for m in return_list],interactive=True) | |
with gr.Row(): | |
stop_button=gr.Button("Stop") | |
clear_btn = gr.Button("Clear") | |
with gr.Row(): | |
with gr.Tab("Text"): | |
data=gr.Textbox(label="Input Data (paste text)", lines=6) | |
with gr.Tab("File"): | |
file=gr.Files(label="Input File(s) (.pdf .txt)") | |
with gr.Tab("Folder"): | |
directory=gr.File(label="Folder", file_count='directory') | |
with gr.Tab("Raw HTML"): | |
url = gr.Textbox(label="URL") | |
with gr.Tab("PDF URL"): | |
pdf_url = gr.Textbox(label="PDF URL") | |
with gr.Tab("PDF Batch"): | |
pdf_batch = gr.Textbox(label="PDF URL Batch (comma separated)") | |
m_box=gr.HTML() | |
zoom_btn=gr.Slider(label="Zoom",step=0.01,minimum=0.1,maximum=20,value=1,interactive=True) | |
e_box=gr.Textbox(interactive=True) | |
with gr.Row(): | |
upd_button=gr.Button("Update Chart") | |
create_im=gr.Button("Create Image") | |
svg_img=gr.Image() | |
url_box=gr.Textbox(interactive=True) | |
json_out=gr.JSON() | |
#text=gr.JSON() | |
#get_score.click(return_score,score,[score],_js=score_js) | |
score=gr.Textbox() | |
def return_score(text): | |
print(text) | |
return text | |
create_im.click(return_score,score,[score],_js=score_js).then(create_image,score,svg_img) | |
#zoom_btn.change(zoom_update,zoom_btn,None) | |
upd_button.click(mm,[e_box,zoom_btn],[m_box,url_box]) | |
#inp_query.change(search_models,inp_query,models_dd) | |
clear_btn.click(clear_fn,None,[prompt,chatbot]) | |
#go=button.click(summarize,[prompt,chatbot,report_check,chart_check,data,file,directory,url,pdf_url,pdf_batch],[prompt,chatbot,e_box,json_out]) | |
go=button.click(summarize,[prompt,chatbot,data,file,directory,url,pdf_url,pdf_batch],[prompt,chatbot,m_box,e_box,json_out,url_box]) | |
stop_button.click(None,None,None,cancels=[go]) | |
#app.queue(default_concurrency_limit=20).launch(show_api=False) | |
app.queue(concurrency_count=20).launch(show_api=False) | |