eliasalbouzidi
commited on
Commit
•
e96ea60
1
Parent(s):
c3a920d
Update README.md
Browse files
README.md
CHANGED
@@ -28,13 +28,13 @@ tags:
|
|
28 |
|
29 |
<!-- Provide a quick summary of what the model is/does. -->
|
30 |
|
31 |
-
This model is designed to categorize text into two classes: "
|
32 |
|
33 |
-
The model was trained using a dataset containing 190,000 labeled text samples, distributed among the two classes of "
|
34 |
|
35 |
The model is based on the Distilbert-base model.
|
36 |
|
37 |
-
In terms of performance, the model has achieved a score of 0.
|
38 |
### Model Description
|
39 |
|
40 |
The model can be used directly to classify text into one of the two classes. It takes in a string of text as input and outputs a probability distribution over the two classes. The class with the highest probability is selected as the predicted class.
|
@@ -49,18 +49,20 @@ The model can be used directly to classify text into one of the two classes. It
|
|
49 |
|
50 |
### Training Procedure
|
51 |
|
52 |
-
The model was trained utilizing the Hugging Face Transformers library. The training approach
|
53 |
|
54 |
### Training Data
|
55 |
|
56 |
-
The training data for the text classification model consists of a large corpus of text labeled with one of the two classes: "
|
57 |
|
58 |
-
100,000 examples labeled as "
|
59 |
|
60 |
-
90,000 examples labeled as "
|
61 |
|
62 |
The data was preprocessed to remove stop words and punctuation, and to convert all text to lowercase.
|
63 |
|
|
|
|
|
64 |
More information about the training data can be found in the Dataset Card (availabe soon).
|
65 |
|
66 |
## Uses
|
|
|
28 |
|
29 |
<!-- Provide a quick summary of what the model is/does. -->
|
30 |
|
31 |
+
This model is designed to categorize text into two classes: "safe", or "nsfw", which makes it suitable for content moderation and filtering applications.
|
32 |
|
33 |
+
The model was trained using a dataset containing 190,000 labeled text samples, distributed among the two classes of "safe" and "nsfw".
|
34 |
|
35 |
The model is based on the Distilbert-base model.
|
36 |
|
37 |
+
In terms of performance, the model has achieved a score of 0.988 for F1.
|
38 |
### Model Description
|
39 |
|
40 |
The model can be used directly to classify text into one of the two classes. It takes in a string of text as input and outputs a probability distribution over the two classes. The class with the highest probability is selected as the predicted class.
|
|
|
49 |
|
50 |
### Training Procedure
|
51 |
|
52 |
+
The model was trained utilizing the Hugging Face Transformers library. The training approach involved first fine-tuning the DistilBERT-base model and then applying transfer learning. Initially, the entire DistilBERT-base model was fine-tuned on the labeled dataset. Following this, transfer learning was employed by freezing the original layers of the fine-tuned DistilBERT model and fine-tuning only the classification layers. This approach allowed the model to leverage the fine-tuned knowledge of the DistilBERT-base model while adapting further to the specific task at hand. To optimize memory usage and accelerate training, mixed precision FP16 was used.
|
53 |
|
54 |
### Training Data
|
55 |
|
56 |
+
The training data for finetuning the text classification model consists of a large corpus of text labeled with one of the two classes: "safe" and "nsfw". The dataset contains a total of 190,000 examples, which are distributed as follows:
|
57 |
|
58 |
+
100,000 examples labeled as "safe"
|
59 |
|
60 |
+
90,000 examples labeled as "nsfw"
|
61 |
|
62 |
The data was preprocessed to remove stop words and punctuation, and to convert all text to lowercase.
|
63 |
|
64 |
+
After fine-tuning the DistilBERT-base model on this dataset, transfer learning was applied using a smaller dataset. For transfer learning, the original layers of the fine-tuned DistilBERT model were frozen, and only the classification layers were fine-tuned on an additional dataset containing 40,000 examples.
|
65 |
+
|
66 |
More information about the training data can be found in the Dataset Card (availabe soon).
|
67 |
|
68 |
## Uses
|