PALIGEMMA DHIVEHI

This model is a fine-tuned version of google/paligemma-3b-pt-224 on an dhivehi-text-image dataset.

Model description

Finetuned version for dhivehi singline image dataset.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 2
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 4
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 2
  • num_epochs: 2

Framework versions

  • PEFT 0.14.0
  • Transformers 4.45.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.0.1
  • Tokenizers 0.20.1
Downloads last month
87
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The HF Inference API does not support image-to-text models for peft library.

Model tree for alakxender/paligemma-dhivehi-xs

Adapter
(171)
this model

Dataset used to train alakxender/paligemma-dhivehi-xs