MahaGemma-2B

MahaGemma-2B is a Marathi Gemma model. It is a Gemma 2B (google/gemma-2b) model LoRA fine-tuned on translated Marathi datasets. [dataset link] (https://github.com/l3cube-pune/MarathiNLP)

This is part of the MahaNLP initiative. More details coming soon.

Prompt format:

<bos>\n### Instruction:\nमहाराष्ट्राची राजधानी काय आहे?\n\n### Input:\n\n\n### Response:\nमहाराष्ट्राची राजधानी मुंबई आहे

Citing

@article{joshi2022l3cube,
  title={L3cube-mahanlp: Marathi natural language processing datasets, models, and library},
  author={Joshi, Raviraj},
  journal={arXiv preprint arXiv:2205.14728},
  year={2022}
}

Model Family:
MahaGemma-2B
MahaGemma-7B

Downloads last month
40
Safetensors
Model size
2.51B params
Tensor type
F32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for l3cube-pune/marathi-gpt-gemma-2b

Quantizations
2 models