sagorsarker commited on
Commit
b0dd619
1 Parent(s): d3d5cdd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -8
README.md CHANGED
@@ -16,7 +16,7 @@ base_model:
16
 
17
  ## Model Information
18
 
19
- This model is a continually pretrained version of the [google/gemma-2-2b](https://huggingface.co/google/gemma-2-2b) architecture, fine-tuned on extensive Bangla datasets. The primary goal of the continual pretraining was to enhance the model's ability to generate high-quality Bangla text. By extending the pretraining process specifically on Bangla data, the model has demonstrated superior performance in tasks related to Bangla language understanding evaluation benchmarks and text generation.
20
 
21
  **Model Architecture:** Gemma 2 is an auto-regressive language model that uses an optimized transformer architecture.
22
 
@@ -31,7 +31,7 @@ Below we share some code snippets on how to get quickly started with running the
31
  pip install -U transformers
32
  ```
33
 
34
- Then, copy the snippet from the section that is relevant for your usecase.
35
 
36
  #### Running with the `pipeline` API
37
 
@@ -62,14 +62,14 @@ print(response)
62
  **Overview:** We have collected a large Bangla raw dataset of text data from a wide variety of sources. Our collected data so far includes a mix of web documents, books, translated text, transliterated text, transcribe text, code-mixed text, conversations, and open sources raw data. The dataset is cleaned and filtered by different filtering criteria to ensure the quality of the data. Our collected data size roughly around 268 GB. We separated __33GB__ data from that using a ratio of the data actual data size. Total trained tokens are __4.4B__ tokens.
63
 
64
  Data sources summary:
65
- - Web documents: Extract, clean, filter common crawl data
66
- - Books: Extract, clean, filter books data
67
  - Transcribed text: Used in-house Bangla ASR model to transcribe Bangla audio data
68
  - Translation data: We trained a Bangla-English translation LLM model and used it to translate English data to Bangla
69
  - Code-mixed data: We trained a Bangla-English code-mixed LLM model and used it to generate code-mixed data
70
  - Transliteration data: We trained a Bangla-English transliteration LLM model and used it to generate transliterated data
71
  - Synthetic data: We generated synthetic data using a Bangla LLM model
72
- - Others: We scrap some selected websites data, used open-sources data, and used some other data sources
73
 
74
 
75
  ## Benchmarks
@@ -77,11 +77,11 @@ Data sources summary:
77
  In this section, we report the results for __titulm-gemma-2-2b-v1.1__ models on standard automatic benchmarks. For all these evaluations, we used [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) evaluations library.
78
 
79
  ### Evaluation Datasets
80
- We evaluated our pretrained models on both Bangla and English benchmark datasets. Although the model is trained on Bangla data, it's English capability is also evaluated on English benchmark datasets. The evaluation datasets are as follows:
81
 
82
  #### Bangla Benchmark datasets
83
  We evaluated the models on the following datasets:
84
- - [Bangla MMLU](): A privated multiple choice questions datasets developed by Hishab curated from various sources.
85
  - [CommonsenseQa Bangla](https://huggingface.co/datasets/hishab/commonsenseqa-bn): A Bangla translation of the CommonsenseQA dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
86
  - [OpenbookQA Bangla](https://huggingface.co/datasets/hishab/openbookqa-bn): A Bangla translation of the OpenbookQA dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
87
  - [Piqa Bangla](https://huggingface.co/datasets/hishab/piqa-bn): A Bangla translation of the Piqa dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
@@ -89,7 +89,7 @@ We evaluated the models on the following datasets:
89
 
90
  #### English Benchmark datasets
91
  - [MMLU](https://huggingface.co/datasets/cais/mmlu): This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge.
92
- - [CommonseQa](https://huggingface.co/datasets/tau/commonsense_qa): CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers .
93
  - [OpenbookQA](https://huggingface.co/datasets/allenai/openbookqa): OpenBookQA aims to promote research in advanced question-answering, probing a deeper understanding of both the topic (with salient facts summarized as an open book, also provided with the dataset) and the language it is expressed in.
94
  - [Piqa](https://huggingface.co/datasets/ybisk/piqa): The PIQA dataset focuses on physical commonsense reasoning, challenging AI to handle everyday situations requiring practical knowledge and unconventional solutions. Inspired by instructables.com, it aims to enhance AI's ability to understand and reason about physical interactions.
95
  - [BoolQ](https://huggingface.co/datasets/google/boolq): BoolQ is a question answering dataset for yes/no questions containing 15942 examples. These questions are naturally occurring ---they are generated in unprompted and unconstrained settings. Each example is a triplet of (question, passage, answer), with the title of the page as optional additional context. The text-pair classification setup is similar to existing natural language inference tasks.
@@ -110,6 +110,15 @@ We evaluated the models on the following datasets:
110
  | | 5-shot | 0.35 | **0.57** | **0.40** | **0.38** | 0.60 |
111
 
112
  #### Evaluation on English Benchmark datasets
 
 
 
 
 
 
 
 
 
113
 
114
  ### Instruction Tuned Models
115
 
 
16
 
17
  ## Model Information
18
 
19
+ This model is a continually pre-trained version of the [google/gemma-2-2b](https://huggingface.co/google/gemma-2-2b) architecture, fine-tuned on extensive Bangla datasets. The primary goal of the continual pretraining was to enhance the model's ability to generate high-quality Bangla text. By extending the pretraining process specifically on Bangla data, the model has demonstrated superior performance in tasks related to Bangla language understanding evaluation benchmarks and text generation.
20
 
21
  **Model Architecture:** Gemma 2 is an auto-regressive language model that uses an optimized transformer architecture.
22
 
 
31
  pip install -U transformers
32
  ```
33
 
34
+ Then, copy the snippet from the section that is relevant to your use case.
35
 
36
  #### Running with the `pipeline` API
37
 
 
62
  **Overview:** We have collected a large Bangla raw dataset of text data from a wide variety of sources. Our collected data so far includes a mix of web documents, books, translated text, transliterated text, transcribe text, code-mixed text, conversations, and open sources raw data. The dataset is cleaned and filtered by different filtering criteria to ensure the quality of the data. Our collected data size roughly around 268 GB. We separated __33GB__ data from that using a ratio of the data actual data size. Total trained tokens are __4.4B__ tokens.
63
 
64
  Data sources summary:
65
+ - Web documents: Extract, clean, and filter common crawl data
66
+ - Books: Extract, clean, and filter book data
67
  - Transcribed text: Used in-house Bangla ASR model to transcribe Bangla audio data
68
  - Translation data: We trained a Bangla-English translation LLM model and used it to translate English data to Bangla
69
  - Code-mixed data: We trained a Bangla-English code-mixed LLM model and used it to generate code-mixed data
70
  - Transliteration data: We trained a Bangla-English transliteration LLM model and used it to generate transliterated data
71
  - Synthetic data: We generated synthetic data using a Bangla LLM model
72
+ - Others: We scrap some selected website data, used open-source data, and used some other data sources
73
 
74
 
75
  ## Benchmarks
 
77
  In this section, we report the results for __titulm-gemma-2-2b-v1.1__ models on standard automatic benchmarks. For all these evaluations, we used [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) evaluations library.
78
 
79
  ### Evaluation Datasets
80
+ We evaluated our pre-trained models on both Bangla and English benchmark datasets. Although the model is trained on Bangla data, it's English capability is also evaluated on English benchmark datasets. The evaluation datasets are as follows:
81
 
82
  #### Bangla Benchmark datasets
83
  We evaluated the models on the following datasets:
84
+ - [Bangla MMLU](): A private multiple choice question dataset developed by Hishab curated from various sources.
85
  - [CommonsenseQa Bangla](https://huggingface.co/datasets/hishab/commonsenseqa-bn): A Bangla translation of the CommonsenseQA dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
86
  - [OpenbookQA Bangla](https://huggingface.co/datasets/hishab/openbookqa-bn): A Bangla translation of the OpenbookQA dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
87
  - [Piqa Bangla](https://huggingface.co/datasets/hishab/piqa-bn): A Bangla translation of the Piqa dataset. The dataset was translated using a new method called Expressive Semantic Translation (EST), which combines Google Machine Translation with LLM-based rewriting modifications.
 
89
 
90
  #### English Benchmark datasets
91
  - [MMLU](https://huggingface.co/datasets/cais/mmlu): This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge.
92
+ - [CommonseQa](https://huggingface.co/datasets/tau/commonsense_qa): CommonsenseQA is a new multiple-choice question-answering dataset that requires different types of commonsense knowledge to predict the correct answers .
93
  - [OpenbookQA](https://huggingface.co/datasets/allenai/openbookqa): OpenBookQA aims to promote research in advanced question-answering, probing a deeper understanding of both the topic (with salient facts summarized as an open book, also provided with the dataset) and the language it is expressed in.
94
  - [Piqa](https://huggingface.co/datasets/ybisk/piqa): The PIQA dataset focuses on physical commonsense reasoning, challenging AI to handle everyday situations requiring practical knowledge and unconventional solutions. Inspired by instructables.com, it aims to enhance AI's ability to understand and reason about physical interactions.
95
  - [BoolQ](https://huggingface.co/datasets/google/boolq): BoolQ is a question answering dataset for yes/no questions containing 15942 examples. These questions are naturally occurring ---they are generated in unprompted and unconstrained settings. Each example is a triplet of (question, passage, answer), with the title of the page as optional additional context. The text-pair classification setup is similar to existing natural language inference tasks.
 
110
  | | 5-shot | 0.35 | **0.57** | **0.40** | **0.38** | 0.60 |
111
 
112
  #### Evaluation on English Benchmark datasets
113
+ - **gemma-2-2b** consistently achieves the highest scores across all tasks in both 0-shot and 5-shot settings, leading in **MMLU**, **BoolQ**, **Commonsense QA**, **OpenBook QA**, and **PIQA**, with a maximum 5-shot score of **0.80** in **PIQA**.
114
+ - **titulm-gemma-2-2b-v1.1** performs well but trails behind **gemma-2-2b**, particularly in **Commonsense QA** and **OpenBook QA**, with the best scores being slightly lower across all tasks.
115
+
116
+ | Model | Shots | MMLU | BoolQ | Commonsense QA | OpenBook QA | PIQA |
117
+ |--------------------------------------|--------|--------------|------------|--------------------|-----------------|-----------|
118
+ | gemma-2-2b | 0-shot | **0.50** | **0.74** | **0.52** | **0.42** | **0.79** |
119
+ | | 5-shot | **0.53** | **0.78** | **0.66** | **0.42** | **0.80** |
120
+ | titulm-gemma-2-2b-v1.1 | 0-shot | 0.40 | 0.71 | 0.37 | 0.36 | 0.76 |
121
+ | | 5-shot | 0.44 | 0.75 | 0.53 | 0.36 | 0.76 |
122
 
123
  ### Instruction Tuned Models
124