--- pipeline_tag: text-classification widget: - text: >- NairiSoft is looking for a highly qualified person with deep knowledge and practical experience in Java programming. The selected candidate will be involved in all stages of the development life cycle. example_title: Current Position Requirments 1 - text: >- Ogma Applications is seeking motivated Senior Developers to work on its worldwide projects. The projects are web applications utilizing latest technologies in video webcasting over internet for web browsers, Televisions and telephone systems. In order to succeed in this team, the incumbent must have the passion and energy to work in an entrepreneurial, and fast paced environment. In addition, the Senior Software Engineer must be an experienced senior architect and technical leader with in-depth knowledge of software development processes. As a senior member of the team in Armenia, Senior Software Engineer will be working closely with other developers and peers in the US and other teams around the globe, to analyze, design, develop, test and deliver the best in class software. example_title: Current Position Requirments 2 - text: >- Armeconombank OJSC is looking for a .Net Developer to join its team. The Software Developer will take part in design and development projects. example_title: Current Position Requirments 3 language: - en tags: - albert - text-classification - recommendation - job - albert-base-v2 - IT --- This repository contains a Albert model designed for text classification. The architecture of the model is based on the Albert Base v2 model. # Library ``` pip install transformers pip install sentencepiece ``` # Example ```python from transformers import AutoModel,AutoTokenizer tokenizer = AutoTokenizer.from_pretrained('Apizhai/Albert-IT-JobRecommendation', use_fast=False), model = AutoModel.from_pretrained('Apizhai/Albert-IT-JobRecommendation') ``` # Training hyperparameters The following hyperparameters were used during training: - max_seq_length: 128 - max_length: 128 - train_batch_size: 4 - eval_batch_size: 32 - num_train_epochs: 10 - evaluate_during_training: False - evaluate_during_training_steps: 100 - use_multiprocessing: False - fp16: True - save_steps: -1 - save_eval_checkpoints: False - save_model_every_epoch: False - no_cache: True - reprocess_input_data: True - overwrite_output_dir: True - preprocess_inputs: False - num_return_sequences: 1 # Score - f1-score: 0.85574 - macro avg: 0.84748 - weighted avg: 0.81575