Spaces:
Runtime error
Runtime error
Commit
·
7147095
1
Parent(s):
b91b62a
chore: update streamlit deprecation
Browse files
app.py
CHANGED
|
@@ -49,8 +49,6 @@ else:
|
|
| 49 |
device = "cpu"
|
| 50 |
max_len = st.sidebar.slider("Max length", 32, 512, 256, 32)
|
| 51 |
beam_size = st.sidebar.slider("Beam size", 1, 10, 3, 1)
|
| 52 |
-
tokenizer = load_tokenizer(model_name)
|
| 53 |
-
model = load_model(model_name, device)
|
| 54 |
|
| 55 |
# create a text input box for each of the following item
|
| 56 |
# CHỈ TIÊU ĐƠN VỊ ĐIỀU KIỆN KPI mục tiêu tháng Tháng 9.2022 Đánh giá T8.2022 So sánh T8.2022 Tăng giảm T9.2021 So sánh T9.2021 Tăng giảm
|
|
@@ -138,13 +136,20 @@ data = {
|
|
| 138 |
"Previous year": previous_year,
|
| 139 |
}
|
| 140 |
|
|
|
|
|
|
|
| 141 |
|
| 142 |
if st.button("Generate"):
|
| 143 |
-
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
)
|
| 149 |
-
|
| 150 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 49 |
device = "cpu"
|
| 50 |
max_len = st.sidebar.slider("Max length", 32, 512, 256, 32)
|
| 51 |
beam_size = st.sidebar.slider("Beam size", 1, 10, 3, 1)
|
|
|
|
|
|
|
| 52 |
|
| 53 |
# create a text input box for each of the following item
|
| 54 |
# CHỈ TIÊU ĐƠN VỊ ĐIỀU KIỆN KPI mục tiêu tháng Tháng 9.2022 Đánh giá T8.2022 So sánh T8.2022 Tăng giảm T9.2021 So sánh T9.2021 Tăng giảm
|
|
|
|
| 136 |
"Previous year": previous_year,
|
| 137 |
}
|
| 138 |
|
| 139 |
+
tokenizer = load_tokenizer(model_name)
|
| 140 |
+
model = load_model(model_name, device)
|
| 141 |
|
| 142 |
if st.button("Generate"):
|
| 143 |
+
if objective_name == "":
|
| 144 |
+
st.error("Please input objective name")
|
| 145 |
+
elif unit == "":
|
| 146 |
+
st.error("Please input unit")
|
| 147 |
+
else:
|
| 148 |
+
with st.spinner("Generating..."):
|
| 149 |
+
input_string = make_input_sentence_from_strings(data)
|
| 150 |
+
print(input_string)
|
| 151 |
+
descriptions = generate_description(
|
| 152 |
+
input_string, model, tokenizer, device, max_len, model_name, beam_size
|
| 153 |
+
)
|
| 154 |
+
|
| 155 |
+
st.success(descriptions)
|
utils.py
CHANGED
|
@@ -15,21 +15,7 @@ import streamlit as st
|
|
| 15 |
from typing import Dict, List
|
| 16 |
|
| 17 |
|
| 18 |
-
|
| 19 |
-
print(f"Using model {args.model_name}")
|
| 20 |
-
model = AutoModelForSeq2SeqLM.from_pretrained(args.model_name)
|
| 21 |
-
model.to(args.device)
|
| 22 |
-
|
| 23 |
-
if args.load_model_path:
|
| 24 |
-
print(f"Loading model from {args.load_model_path}")
|
| 25 |
-
model.load_state_dict(
|
| 26 |
-
torch.load(args.load_model_path, map_location=torch.device(args.device))
|
| 27 |
-
)
|
| 28 |
-
|
| 29 |
-
return model
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
@st.cache(allow_output_mutation=True)
|
| 33 |
def load_model(model_name, device):
|
| 34 |
print(f"Using model {model_name}")
|
| 35 |
os.makedirs("cache", exist_ok=True)
|
|
@@ -46,7 +32,7 @@ def load_model(model_name, device):
|
|
| 46 |
return model
|
| 47 |
|
| 48 |
|
| 49 |
-
@st.
|
| 50 |
def load_tokenizer(model_name):
|
| 51 |
print(f"Loading tokenizer {model_name}")
|
| 52 |
if "mbart" in model_name.lower():
|
|
|
|
| 15 |
from typing import Dict, List
|
| 16 |
|
| 17 |
|
| 18 |
+
@st.cache_resource
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
def load_model(model_name, device):
|
| 20 |
print(f"Using model {model_name}")
|
| 21 |
os.makedirs("cache", exist_ok=True)
|
|
|
|
| 32 |
return model
|
| 33 |
|
| 34 |
|
| 35 |
+
@st.cache_resource
|
| 36 |
def load_tokenizer(model_name):
|
| 37 |
print(f"Loading tokenizer {model_name}")
|
| 38 |
if "mbart" in model_name.lower():
|