--- license: apache-2.0 language: - fr - en base_model: - google-bert/bert-base-uncased pipeline_tag: text-classification library_name: sentence-transformers --- # Takeda Section Classifier Pretrained model (finetuned version of [BERT Multilingual Uncased](https://huggingface.co/google-bert/bert-base-multilingual-uncased)) on french and english documents using supervised training for sections classification. This work has been made by Digital Innovation Team from Belgium 🇧🇪 (LE). ## Model Description The model aims at classifying text in classes representing part of reports: * Description * Immediate Correction * Root Cause * Action Plan * Impacted Elements ## Intended uses & limitations The model can be use for Takeda documentation, the team do not guarantee results for out of the scope documentation. ## How to Use You can use this model directly with a pipeline for text classification: ```python from transformers import ( TextClassificationPipeline, AutoTokenizer, AutoModelForSequenceClassification, ) tokenizer = AutoTokenizer.from_pretrained("TakedaAIML/section_classifier") model = AutoModelForSequenceClassification.from_pretrained( "TakedaAIML/section_classifier" ) pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer) prediction = pipe('this is a piece of text representing the Description section. An event occur on june 24 and ...') ```