Rename README.md to !huggingface-cli login

#91
No description provided.

Import necessary classes for model loading and quantization

from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig

Configure model quantization to 4-bit for memory and computation efficiency

quantization_config = BitsAndBytesConfig(load_in_4bit=True)

Load the tokenizer for the Gemma 7B Italian model

tokenizer = AutoTokenizer.from_pretrained("google/gemma-7b-it")

Load the Gemma 7B Italian model itself, with 4-bit quantization

model = AutoModelForCausalLM.from_pretrained("google/gemma-7b-it",
quantization_config=quantization_config)

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment