tedgwara's picture
Update app.py
4e48a17
raw
history blame
4.16 kB
import gradio as gr
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LogisticRegression
from sklearn import metrics
from reader import get_article
### ------------------------------ ###
### data transformation ###
### ------------------------------ ###
# options constants
options = [
['Very Poorly Aligned', 'Poorly Aligned', 'Neutrally Aligned', 'Well Aligned', 'Very Well Aligned'],
['Very Limited Experience', 'Limited Experience', 'Neutral Experience', 'Extensive Experience', 'Very Extensive Experience'],
['Extremely Unattractive', 'Moderately Unattractive', 'Neutrally Attractive', 'Moderately Attractive', 'Extremely Attractive'],
['Very Unfavorable', 'Moderately Unfavorable', 'Neutrally Favorable', 'Moderately Favorable', 'Very Favorable'],
['Very Poor Fit', 'Poor Fit', 'Neutral Fit', 'Moderately Good Fit', 'Excellent Fit']
]
# load dataset
uncleaned_data = pd.read_csv('data.csv')
data = pd.DataFrame()
# keep track of which columns are categorical and what
# those columns' value mappings are
# structure: {colname1: {...}, colname2: {...} }
cat_value_dicts = {}
col = 0
final_colname = uncleaned_data.columns[4]
# for each column...
for (colname, colval) in uncleaned_data.iteritems():
# structure: {0: "lilac", 1: "blue", ...}
new_dict = {}
transformed_col_vals = [] # new numeric datapoints
# if not, for each item in that column...
for (row, item) in enumerate(colval.values):
# if item is not in this col's dict...
if item not in new_dict:
new_dict[item] = options[col].index(item)
# then add numerical value to transformed dataframe
transformed_col_vals.append(new_dict[item])
# reverse dictionary only for final col (0, 1) => (vals)
if colname == final_colname:
new_dict = {value : key for (key, value) in new_dict.items()}
cat_value_dicts[colname] = new_dict
data[colname] = transformed_col_vals
col += 1
### -------------------------------- ###
### model training ###
### -------------------------------- ###
# select features and predicton; automatically selects last column as prediction
num_features = 4
x = data.iloc[: , :num_features]
y = data.iloc[: , num_features:]
# split data into training and testing sets
x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.25)
# instantiate the model (using default parameters)
model = LogisticRegression(max_iter=100)
model.fit(x_train, y_train.values.ravel())
y_pred = model.predict(x_test)
### -------------------------------- ###
### article generation ###
### -------------------------------- ###
# borrow file reading function from reader.py
def get_feat():
feats = [abs(x) for x in model.coef_[0]]
max_val = max(feats)
idx = feats.index(max_val)
return data.columns[idx]
acc = str(round(metrics.accuracy_score(y_test, y_pred) * 100, 2)) + '%'
feat = get_feat()
info = get_article(acc, feat)
### ------------------------------- ###
### interface creation ###
### ------------------------------- ###
def predictor(*args):
features = []
# transform categorical input
for num, col in enumerate(args):
features.append(cat_value_dicts[data.columns[num]][col])
# predict single datapoint
new_input = [features]
result = model.predict(new_input)
return cat_value_dicts[final_colname][result[0]]
# add data labels to replace those lost via star-args
inputls = []
labels = [
"How Well Do They Align with RS21's 9 Core Values?",
"How Experienced Are They in RS21's Markets?",
"How Attractive is Their Valuation of RS21?",
"How Favorable is Their Proposed Deal Structure for RS21?"
]
for num, colname in enumerate(labels):
# access categories dict if data is categorical
inputls.append(gr.inputs.Radio(choices=options[num], type="value", label=labels[num]))
# generate gradio interface
interface = gr.Interface(predictor, inputs=inputls, outputs="text", article=info['article'], css=info['css'], theme="grass", title=info['title'], allow_flagging='never', description=info['description'])
# show the interface
interface.launch()