Spaces:
Running
on
Zero
Running
on
Zero
NGUYEN, Xuan Phi
commited on
Commit
•
5100e68
1
Parent(s):
e9cbae4
update
Browse files
app.py
CHANGED
@@ -666,14 +666,12 @@ def debug_chat_response_echo(
|
|
666 |
|
667 |
|
668 |
# ============ CONSTANT ============
|
|
|
669 |
MODEL_TITLE = "DAMO-SeaL-13B - An Assistant for South East Asian Languages"
|
670 |
MODEL_DESC = """
|
671 |
This is a 13B DAMO-SeaL-Chat assistant model built by DAMO Academy, Alibaba Group. It can produce helpful responses in English, Vietnamese, Indonesian and Thai.
|
672 |
-
|
673 |
-
|
674 |
-
|
675 |
-
cite_markdown = """
|
676 |
-
## Citation
|
677 |
If you find our project useful, hope you can star our repo and cite our paper as follows:
|
678 |
```
|
679 |
@article{damonlpsg2023seallm,
|
@@ -681,6 +679,11 @@ If you find our project useful, hope you can star our repo and cite our paper as
|
|
681 |
title = {SeaL: A language model for South East Asian Languages},
|
682 |
year = 2023,
|
683 |
}
|
|
|
|
|
|
|
|
|
|
|
684 |
"""
|
685 |
# journal = {arXiv preprint arXiv:2306.02858}
|
686 |
# url = {https://arxiv.org/abs/2306.02858}
|
@@ -699,7 +702,6 @@ def launch():
|
|
699 |
global demo, llm, DEBUG
|
700 |
model_desc = MODEL_DESC
|
701 |
model_path = MODEL_PATH
|
702 |
-
assert os.path.exists(model_path), f'{model_path} not found'
|
703 |
model_title = MODEL_TITLE
|
704 |
tensor_parallel = TENSOR_PARALLEL
|
705 |
assert tensor_parallel > 0 , f'{tensor_parallel} invalid'
|
@@ -708,10 +710,11 @@ def launch():
|
|
708 |
max_tokens = 4096
|
709 |
|
710 |
if DEBUG:
|
711 |
-
model_desc += "<br>!!!!! This is in debug mode, responses will be copy original"
|
712 |
response_fn = debug_chat_response_echo
|
713 |
else:
|
714 |
# ! load the model
|
|
|
715 |
llm = LLM(model=model_path, dtype=dtype, tensor_parallel_size=tensor_parallel)
|
716 |
|
717 |
print(f'Use system prompt:\n{sys_prompt}')
|
@@ -723,6 +726,7 @@ def launch():
|
|
723 |
demo = gr.ChatInterface(
|
724 |
response_fn,
|
725 |
chatbot=ChatBot(
|
|
|
726 |
bubble_full_width=False,
|
727 |
latex_delimiters=[
|
728 |
{ "left": "$", "right": "$", "display": False},
|
@@ -742,7 +746,32 @@ def launch():
|
|
742 |
gr.Textbox(value=sys_prompt, label='System prompt', lines=8)],
|
743 |
)
|
744 |
|
745 |
-
gr.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
746 |
demo.queue()
|
747 |
# demo.launch(server_port=args.port)
|
748 |
demo.launch()
|
|
|
666 |
|
667 |
|
668 |
# ============ CONSTANT ============
|
669 |
+
MODEL_NAME = "DAMO-SeaL-13B"
|
670 |
MODEL_TITLE = "DAMO-SeaL-13B - An Assistant for South East Asian Languages"
|
671 |
MODEL_DESC = """
|
672 |
This is a 13B DAMO-SeaL-Chat assistant model built by DAMO Academy, Alibaba Group. It can produce helpful responses in English, Vietnamese, Indonesian and Thai.
|
673 |
+
<br>
|
674 |
+
#### Citation
|
|
|
|
|
|
|
675 |
If you find our project useful, hope you can star our repo and cite our paper as follows:
|
676 |
```
|
677 |
@article{damonlpsg2023seallm,
|
|
|
679 |
title = {SeaL: A language model for South East Asian Languages},
|
680 |
year = 2023,
|
681 |
}
|
682 |
+
```
|
683 |
+
""".strip()
|
684 |
+
|
685 |
+
|
686 |
+
cite_markdown = """
|
687 |
"""
|
688 |
# journal = {arXiv preprint arXiv:2306.02858}
|
689 |
# url = {https://arxiv.org/abs/2306.02858}
|
|
|
702 |
global demo, llm, DEBUG
|
703 |
model_desc = MODEL_DESC
|
704 |
model_path = MODEL_PATH
|
|
|
705 |
model_title = MODEL_TITLE
|
706 |
tensor_parallel = TENSOR_PARALLEL
|
707 |
assert tensor_parallel > 0 , f'{tensor_parallel} invalid'
|
|
|
710 |
max_tokens = 4096
|
711 |
|
712 |
if DEBUG:
|
713 |
+
model_desc += "\n<br>!!!!! This is in debug mode, responses will be copy original"
|
714 |
response_fn = debug_chat_response_echo
|
715 |
else:
|
716 |
# ! load the model
|
717 |
+
assert os.path.exists(model_path), f'{model_path} not found'
|
718 |
llm = LLM(model=model_path, dtype=dtype, tensor_parallel_size=tensor_parallel)
|
719 |
|
720 |
print(f'Use system prompt:\n{sys_prompt}')
|
|
|
726 |
demo = gr.ChatInterface(
|
727 |
response_fn,
|
728 |
chatbot=ChatBot(
|
729 |
+
# value=MODEL_NAME,
|
730 |
bubble_full_width=False,
|
731 |
latex_delimiters=[
|
732 |
{ "left": "$", "right": "$", "display": False},
|
|
|
746 |
gr.Textbox(value=sys_prompt, label='System prompt', lines=8)],
|
747 |
)
|
748 |
|
749 |
+
# with gr.Blocks() as demo:
|
750 |
+
# gr.ChatInterface(
|
751 |
+
# response_fn,
|
752 |
+
# chatbot=ChatBot(
|
753 |
+
# bubble_full_width=False,
|
754 |
+
# latex_delimiters=[
|
755 |
+
# { "left": "$", "right": "$", "display": False},
|
756 |
+
# { "left": "$$", "right": "$$", "display": True},
|
757 |
+
# ]
|
758 |
+
# ),
|
759 |
+
# textbox=gr.Textbox(placeholder='Type message', lines=8, max_lines=128, min_width=200),
|
760 |
+
# submit_btn=gr.Button(value='Submit', variant="primary", scale=0),
|
761 |
+
# # stop_btn=None,
|
762 |
+
# title=f"{model_title}",
|
763 |
+
# description=f"{model_desc}",
|
764 |
+
# # ! decide if can change the system prompt.
|
765 |
+
# additional_inputs=[
|
766 |
+
# gr.Number(value=0, label='Temperature (higher -> more random)'),
|
767 |
+
# gr.Number(value=max_tokens, label='Max generated tokens (increase if want more generation)'),
|
768 |
+
# gr.Number(value=0.4, label='Frequency penalty (> 0 encourage new tokens)'),
|
769 |
+
# gr.Textbox(value=sys_prompt, label='System prompt', lines=8)
|
770 |
+
# ],
|
771 |
+
# )
|
772 |
+
|
773 |
+
# gr.Markdown(cite_markdown)
|
774 |
+
|
775 |
demo.queue()
|
776 |
# demo.launch(server_port=args.port)
|
777 |
demo.launch()
|