smartPatent-mContriever-lora
The model is fine-tuned on the customed Korean Patent Retrieval system.
Training Data
Two types of datasets are used as training data: queries automatically generated through GPT-4 and patent titles that are linked to existing patent abstracts.
Usage
from transformers import AutoTokenizer, AutoModel, AutoModelForSequenceClassification
import torch
from transformers import AutoModel, AutoTokenizer
from peft import PeftModel, PeftConfig
def get_model(peft_model_name):
config = PeftConfig.from_pretrained(peft_model_name)
base_model = AutoModel.from_pretrained(config.base_model_name_or_path)
model = PeftModel.from_pretrained(base_model, peft_model_name)
model = model.merge_and_unload()
model.eval()
return model
# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained('facebook/mcontriever-msmarco')
model = get_model('hanseokOh/smartPatent-mContriever-lora')
Info
- Developed by: hanseokOh
- Model type: information retriever
- Language(s) (NLP): Korean
- Finetuned from model [optional]: mContriever-msmarco
Model Sources [optional]
- Repository: https://github.com/hanseokOh/PatentSearch
- Downloads last month
- 5
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no pipeline_tag.
Model tree for hanseokOh/smartPatent-mContriever-lora
Base model
facebook/mcontriever-msmarco