How to run the model in oobabooga-text-generation-webui
Hi,
I run the models in oobabooga-text-generation-webui, and got below error:
Anyone know how to solve? Thank you.
Traceback (most recent call last):
File "D:\AI\oobabooga_windows\text-generation-webui\server.py", line 68, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 78, in load_model
output = load_func_maploader
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 148, in huggingface_loader
model = LoaderClass.from_pretrained(Path(f"{shared.args.model_dir}/{model_name}"), low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if shared.args.bf16 else torch.float16, trust_remote_code=shared.args.trust_remote_code)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\models\auto\auto_factory.py", line 480, in from_pretrained
model_class = get_class_from_dynamic_module(
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 431, in get_class_from_dynamic_module
final_module = get_cached_module_file(
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 268, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 151, in check_imports
raise ImportError(
ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run pip install transformers_stream_generator
pip install transformers_stream_generator呢
pip install transformers_stream_generator呢
谢谢。
试过了,没用。
是要进入什么虚拟环境安装,还是base里安装?
Can try to copy file into \installer_files\env\Lib\site-packages but I encounted another issue as below
“.cache\huggingface\modules\transformers_modules\Qwen_Qwen-7B\tokenization_qwen.py”, line 38, in _load_tiktoken_bpe with open(tiktoken_bpe_file, “rb”) as f: TypeError: expected str, bytes or os.PathLike object, not NoneType
Can try to copy file into \installer_files\env\Lib\site-packages but I encounted another issue as below
“.cache\huggingface\modules\transformers_modules\Qwen_Qwen-7B\tokenization_qwen.py”, line 38, in _load_tiktoken_bpe with open(tiktoken_bpe_file, “rb”) as f: TypeError: expected str, bytes or os.PathLike object, not NoneType
这个问题是textgen-webui自动下载模型文件有问题。
用git自己下载模型即可:
git lfs install
git clone https://huggingface.co/Qwen/Qwen-7B
Can try to copy file into \installer_files\env\Lib\site-packages but I encounted another issue as below
“.cache\huggingface\modules\transformers_modules\Qwen_Qwen-7B\tokenization_qwen.py”, line 38, in _load_tiktoken_bpe with open(tiktoken_bpe_file, “rb”) as f: TypeError: expected str, bytes or os.PathLike object, not NoneType这个问题是textgen-webui自动下载模型文件有问题。
用git自己下载模型即可:git lfs install git clone https://huggingface.co/Qwen/Qwen-7B
确实是自己下载的模型不过好像还是不行
Can try to copy file into \installer_files\env\Lib\site-packages but I encounted another issue as below
“.cache\huggingface\modules\transformers_modules\Qwen_Qwen-7B\tokenization_qwen.py”, line 38, in _load_tiktoken_bpe with open(tiktoken_bpe_file, “rb”) as f: TypeError: expected str, bytes or os.PathLike object, not NoneType这个问题是textgen-webui自动下载模型文件有问题。
用git自己下载模型即可:git lfs install git clone https://huggingface.co/Qwen/Qwen-7B
谢谢,不过好像还是不行。
Traceback (most recent call last):
File "D:\AI\oobabooga_windows\text-generation-webui\modules\ui_model_menu.py", line 185, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 79, in load_model
output = load_func_maploader
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 149, in huggingface_loader
model = LoaderClass.from_pretrained(Path(f"{shared.args.model_dir}/{model_name}"), low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if shared.args.bf16 else torch.float16, trust_remote_code=shared.args.trust_remote_code)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\models\auto\auto_factory.py", line 498, in from_pretrained
model_class = get_class_from_dynamic_module(
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 451, in get_class_from_dynamic_module
final_module = get_cached_module_file(
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 279, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\dynamic_module_utils.py", line 152, in check_imports
raise ImportError(
ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run pip install transformers_stream_generator
- 进入oobabooga_windows文件夹
- 运行cmd_windows.bat
- 输入 pip install transformers_stream_generator
- 输入 pip install tiktoken
- 在web ui 界面中的model页面左侧,勾选trust-remote-code后,load model
it reports error
File "Qwen-7B/modeling_qwen.py", line 1111, in generate
return super().generate(
File "/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "lib/python3.10/site-packages/transformers/generation/utils.py", line 1296, in generate
eos_token_id = eos_token_id[0]
IndexError: list index out of range
Output generated in 0.48 seconds (0.00 tokens/s, 0 tokens, context 61, seed 1066529497)
- 进入oobabooga_windows文件夹
- 运行cmd_windows.bat
- 输入 pip install transformers_stream_generator
- 输入 pip install tiktoken
- 在web ui 界面中的model页面左侧,勾选trust-remote-code后,load model
谢谢
我将transformers_stream_generator拷贝到.\oobabooga_windows\installer_files\env\Lib\site-packages里,
出现下面这个新报错,请问怎么解? :)
Traceback (most recent call last):
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\modeling_utils.py", line 479, in load_state_dict
return torch.load(checkpoint_file, map_location=map_location)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\torch\serialization.py", line 797, in load
with _open_zipfile_reader(opened_file) as opened_zipfile:
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\torch\serialization.py", line 283, in init
super().init(torch._C.PyTorchFileReader(name_or_buffer))
RuntimeError: PytorchStreamReader failed reading zip archive: failed finding central directory
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\modeling_utils.py", line 483, in load_state_dict
if f.read(7) == "version":
UnicodeDecodeError: 'gbk' codec can't decode byte 0x80 in position 64: illegal multibyte sequence
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\AI\oobabooga_windows\text-generation-webui\modules\ui_model_menu.py", line 185, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 79, in load_model
output = load_func_maploader
File "D:\AI\oobabooga_windows\text-generation-webui\modules\models.py", line 149, in huggingface_loader
model = LoaderClass.from_pretrained(Path(f"{shared.args.model_dir}/{model_name}"), low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if shared.args.bf16 else torch.float16, trust_remote_code=shared.args.trust_remote_code)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\models\auto\auto_factory.py", line 511, in from_pretrained
return model_class.from_pretrained(
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\modeling_utils.py", line 2805, in from_pretrained
state_dict = load_state_dict(resolved_archive_file)
File "D:\AI\oobabooga_windows\installer_files\env\lib\site-packages\transformers\modeling_utils.py", line 495, in load_state_dict
raise OSError(
OSError: Unable to load weights from pytorch checkpoint file for 'models\Qwen-7B\pytorch_model.bin' at 'models\Qwen-7B\pytorch_model.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.
检查下preset是不是用的mirostat,这个需要llama.cpp加载才能用,其他加载器好像就会发疯、
谢谢,用了mirostat和llama.cpp加载,确实好了不少,但还是经常胡说八道,哈哈。
hhh,但是我的意思是,尽量别用mirostat……mirostat按reddit社区的测试,适合一些模型+long relpies用来长文回复(例如用Hermes做R18场景长文描述),一般情况下oobabooga推荐的是这么几个配置
Instruction following:
Divine Intellect
Big O
simple-1
Chat:
Midnight Enigma
Yara
Shortwave
hhh,但是我的意思是,尽量别用mirostat……mirostat按reddit社区的测试,适合一些模型+long relpies用来长文回复(例如用Hermes做R18场景长文描述),一般情况下oobabooga推荐的是这么几个配置
Instruction following:
Divine Intellect
Big O
simple-1Chat:
Midnight Enigma
Yara
Shortwave
谢谢。
请问,Filter by loader里要选吗?还是保持All就行?
Help me guys,issues below:
Traceback (most recent call last):
File “D:\Downloads\text-generation-webui-main\modules\ui_model_menu.py”, line 201, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File “D:\Downloads\text-generation-webui-main\modules\models.py”, line 78, in load_model
output = load_func_maploader
File “D:\Downloads\text-generation-webui-main\modules\models.py”, line 122, in huggingface_loader
config = AutoConfig.from_pretrained(path_to_model, trust_remote_code=params['trust_remote_code'])
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\models\auto\configuration_auto.py”, line 1037, in from_pretrained
config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\configuration_utils.py”, line 620, in get_config_dict
config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\configuration_utils.py”, line 675, in _get_config_dict
resolved_config_file = cached_file(
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\utils\hub.py”, line 400, in cached_file
raise EnvironmentError(
OSError: models\Qwen-7B does not appear to have a file named config.json. Checkout ‘https://huggingface.co/models\Qwen-7B/None’ for available files.
请参考下这几个问题的回复
https://huggingface.co/Qwen/Qwen-7B-Chat/discussions/26#64d306b50f17d186414fe550
https://github.com/QwenLM/Qwen/issues/361
如果是预训练模型的话,没有经过对话式的人类对齐微调,对话模式不是很稳定(即使用了preset)。对话建议使用Chat模型。
Chat模型的输入是ChatML格式的,text-generation-webui用文本生成模式调用Chat模型的话,Text generation
那个tab里建议这么填写(不确定它是不是又改版了)
<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
这里填内容<|im_end|>
<|im_start|>assistant
而且text-generation-webui的默认配置与Qwen并不兼容,它是看输出字符串看是否有中止字符串来判断中止的,需要设置skip_special_tokens
为False
,且配置custom_stopping_strings
为"<|im_end|>", "<|im_start|>", "<|endoftext|>"
,这些可以在Parameters
那个tab里设置。
Help me guys,issues below:
Traceback (most recent call last):File “D:\Downloads\text-generation-webui-main\modules\ui_model_menu.py”, line 201, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File “D:\Downloads\text-generation-webui-main\modules\models.py”, line 78, in load_modeloutput = load_func_maploader
File “D:\Downloads\text-generation-webui-main\modules\models.py”, line 122, in huggingface_loaderconfig = AutoConfig.from_pretrained(path_to_model, trust_remote_code=params['trust_remote_code'])
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\models\auto\configuration_auto.py”, line 1037, in from_pretrainedconfig_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\configuration_utils.py”, line 620, in get_config_dictconfig_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\configuration_utils.py”, line 675, in _get_config_dictresolved_config_file = cached_file(
File “D:\Downloads\text-generation-webui-main\installer_files\env\lib\site-packages\transformers\utils\hub.py”, line 400, in cached_fileraise EnvironmentError(
OSError: models\Qwen-7B does not appear to have a file named config.json. Checkout ‘https://huggingface.co/models\Qwen-7B/None’ for available files.
It seems that the file failed to download. Please check if the files were there. You can try download and place the files manually and see if the instructions from otgw can help you. https://github.com/oobabooga/text-generation-webui#downloading-models
请参考下这几个问题的回复
https://huggingface.co/Qwen/Qwen-7B-Chat/discussions/26#64d306b50f17d186414fe550
https://github.com/QwenLM/Qwen/issues/361如果是预训练模型的话,没有经过对话式的人类对齐微调,对话模式不是很稳定(即使用了preset)。对话建议使用Chat模型。
Chat模型的输入是ChatML格式的,text-generation-webui用文本生成模式调用Chat模型的话,
Text generation
那个tab里建议这么填写(不确定它是不是又改版了)<|im_start|>system You are a helpful assistant.<|im_end|> <|im_start|>user 这里填内容<|im_end|> <|im_start|>assistant
而且text-generation-webui的默认配置与Qwen并不兼容,它是看输出字符串看是否有中止字符串来判断中止的,需要设置
skip_special_tokens
为False
,且配置custom_stopping_strings
为"<|im_end|>", "<|im_start|>", "<|endoftext|>"
,这些可以在Parameters
那个tab里设置。
谢谢,这个回复到位。
怎么设置text-generation-webui,才能让Qwen处理8万汉字以上的文本?