Apizhai's picture
Update README.md
eaf1120
|
raw
history blame
2.57 kB
---
pipeline_tag: text-classification
widget:
- text: >-
NairiSoft is looking for a highly qualified person with deep knowledge and
practical experience in Java programming. The selected candidate will be
involved in all stages of the development life cycle.
example_title: Current Position Requirments 1
- text: >-
Ogma Applications is seeking motivated Senior Developers to work on its
worldwide projects. The projects are web applications utilizing latest
technologies in video webcasting over internet for web browsers, Televisions
and telephone systems. In order to succeed in this team, the incumbent must
have the passion and energy to work in an entrepreneurial, and fast paced
environment. In addition, the Senior Software Engineer must be an
experienced senior architect and technical leader with in-depth knowledge of
software development processes. As a senior member of the team in Armenia,
Senior Software Engineer will be working closely with other developers and
peers in the US and other teams around the globe, to analyze, design,
develop, test and deliver the best in class software.
example_title: Current Position Requirments 2
- text: >-
Armeconombank OJSC is looking for a .Net Developer to join its team. The
Software Developer will take part in design and development projects.
example_title: Current Position Requirments 3
language:
- en
tags:
- albert
- text-classification
- recommendation
- job
- albert-base-v2
- IT
---
This repository contains a Albert model designed for text classification. The architecture of the model is based on the Albert Base v2 model.
# Library
```
pip install transformers
pip install sentencepiece
```
# Example
```python
from transformers import AutoModel,AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained('Apizhai/Albert-IT-JobRecommendation', use_fast=False),
model = AutoModel.from_pretrained('Apizhai/Albert-IT-JobRecommendation')
```
# Training hyperparameters
The following hyperparameters were used during training:
- max_seq_length: 128
- max_length: 128
- train_batch_size: 4
- eval_batch_size: 32
- num_train_epochs: 10
- evaluate_during_training: False
- evaluate_during_training_steps: 100
- use_multiprocessing: False
- fp16: True
- save_steps: -1
- save_eval_checkpoints: False
- save_model_every_epoch: False
- no_cache: True
- reprocess_input_data: True
- overwrite_output_dir: True
- preprocess_inputs: False
- num_return_sequences: 1
# Score
- f1-score: 0.85574
- macro avg: 0.84748
- weighted avg: 0.81575