Jeffrey Rathgeber Jr commited on
Commit
c39351b
·
unverified ·
1 Parent(s): 0a5c3d5
Files changed (1) hide show
  1. app.py +7 -0
app.py CHANGED
@@ -57,6 +57,13 @@ if option == 'MILESTONE 3':
57
  tokenizers = [tokenizer_0, tokenizer_1, tokenizer_2, tokenizer_3, tokenizer_4, tokenizer_5]
58
  classifiers = [classifier_0, classifier_1, classifier_2, classifier_3, classifier_4, classifier_5]
59
 
 
 
 
 
 
 
 
60
  # X_train = [textIn]
61
  # batch = tokenizer_0(X_train, padding=True, truncation=True, max_length=512, return_tensors="pt")
62
 
 
57
  tokenizers = [tokenizer_0, tokenizer_1, tokenizer_2, tokenizer_3, tokenizer_4, tokenizer_5]
58
  classifiers = [classifier_0, classifier_1, classifier_2, classifier_3, classifier_4, classifier_5]
59
 
60
+ st.write('IF YOURE READING THIS: I was unable to complete a fully functioning milestone 3. \
61
+ If this message print, that means my program successfully loaded my pretrained models. \
62
+ They are fine tuned iterations of the Bert uncased model, trained on the given training data. \
63
+ The issue arose when I tried to use my models to analyze the input string, and after much troubleshooting, \
64
+ I was unable to get it to work. My pre-processing and training algorithm, along with each models .json and config \
65
+ files will be linked in the github along with the tokenizer I used.')
66
+
67
  # X_train = [textIn]
68
  # batch = tokenizer_0(X_train, padding=True, truncation=True, max_length=512, return_tensors="pt")
69