iqrakiran commited on
Commit
219072f
1 Parent(s): 2bc3018

Add new SentenceTransformer model.

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,440 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilbert/distilroberta-base
3
+ datasets: []
4
+ language: []
5
+ library_name: sentence-transformers
6
+ pipeline_tag: sentence-similarity
7
+ tags:
8
+ - sentence-transformers
9
+ - sentence-similarity
10
+ - feature-extraction
11
+ - generated_from_trainer
12
+ - dataset_size:4093
13
+ - loss:TripletLoss
14
+ - loss:ContrastiveLoss
15
+ - loss:CosineSimilarityLoss
16
+ - loss:MultipleNegativesRankingLoss
17
+ widget:
18
+ - source_sentence: I am angry about an incident at the supermarket, and it has been
19
+ bothering me for the last three or four days. The person involved was flirting
20
+ with me, and I think they should apologize. I didn't ask for an apology or indicate
21
+ that I was upset during the conversation because I felt it would be inappropriate.
22
+ I believe they should call me and set aside an appropriate time to apologize.
23
+ sentences:
24
+ - If you are a counsellor, please answer the questions based on the description
25
+ of the patient.
26
+ - It seems like this incident really impacted you and you believe that an apology
27
+ is necessary for closure. It's understandable that you didn't address it on the
28
+ spot if you felt it was inappropriate. However, people sometimes don't realize
29
+ they've upset someone unless it's brought to their attention. Since you feel so
30
+ strongly about this, have you considered reaching out to the person to express
31
+ your feelings and provide them an opportunity to apologize?
32
+ - It sounds like you have built a strong support system in San Diego with meaningful
33
+ relationships and emotional investments. It's natural to feel apprehensive about
34
+ moving away from that. Maybe you could consider the potential opportunities that
35
+ LA has to offer and weigh them against what you would be leaving behind in San
36
+ Diego. Additionally, think about the ways you can stay connected with your loved
37
+ ones after moving. Just remember that it's a difficult decision, and it's important
38
+ to give yourself time to make the best choice for you.
39
+ - source_sentence: Lately, I've been feeling a bit overwhelmed with everything going
40
+ on in my life. I've been hearing things that aren't really there, like whispers
41
+ and someone calling my name, but it's not happening all the time, just occasionally.
42
+ I also find myself getting a bit disorganized, but I wouldn't say I'm completely
43
+ disorganized. Sometimes, I just feel disconnected with what is happening around
44
+ me. I've been like this for a couple of weeks now. My parents said that I've been
45
+ acting differently, but I don't know, maybe I'm just overreacting? Can you help
46
+ me with this?
47
+ sentences:
48
+ - If you are a counsellor, please answer the questions based on the description
49
+ of the patient.
50
+ - Thank you for sharing your experience with me. I can understand how hearing things
51
+ and feeling disorganized might be confusing and concerning for you. It's important
52
+ to discuss these symptoms and feelings in order to provide proper support. Can
53
+ you tell me if you've experienced any changes in your mood, daily activities,
54
+ or sleep habits during this time? Also, have you noticed any patterns or triggers
55
+ for when you hear the whispers or feel disconnected from what's happening around
56
+ you?
57
+ - I appreciate you sharing your feelings and experiences with me. It sounds like
58
+ you're going through a challenging period in your life, dealing with stress and
59
+ worry. It's important to understand the factors contributing to your stress and
60
+ how they might be influencing your thoughts and emotions. Could you provide some
61
+ more specific examples of the situations that have caused you to feel stressed
62
+ and worried? Additionally, have you noticed any patterns or triggers that seem
63
+ to make your worries more intense?
64
+ - source_sentence: I've been experiencing anxiety due to the difficult transition
65
+ from being at home, taking care of my son, to moving back into the work world.
66
+ I've been trying to manage this anxiety by paying attention to nutrition, exercise,
67
+ and practicing gratitude. Although I feel like I'm managing better now than when
68
+ I first started, I'm still struggling.
69
+ sentences:
70
+ - It's great to hear that you've made some progress in managing your anxiety through
71
+ healthy practices like focusing on nutrition, exercise, and gratitude. It's common
72
+ for individuals to face challenges while transitioning between roles, and it might
73
+ take more time for you to adjust fully. If you still find yourself struggling,
74
+ consider seeking additional support through therapy or support groups. Is there
75
+ a specific aspect of this transition that continues to cause you worry?
76
+ - If you are a counsellor, please answer the questions based on the description
77
+ of the patient.
78
+ - It's great to hear that you have a strong drive for self-improvement and setting
79
+ ambitious goals for yourself. Acknowledging this inner motivation is an essential
80
+ step towards achieving what you desire. Remember, it's important to embrace the
81
+ process of learning and growing, and not to put too much pressure on yourself
82
+ to reach an end goal quickly. Take it one step at a time, and don't forget to
83
+ celebrate your milestones along the way.
84
+ - source_sentence: Recently, I have been experiencing difficulties while learning.
85
+ Sometimes it's quite challenging, and I feel upset with myself to the point of
86
+ crying. However, at other times, I find it easy and feel a sense of accomplishment
87
+ when I finally get things right.
88
+ sentences:
89
+ - If you are a counsellor, please answer the questions based on the description
90
+ of the patient.
91
+ - Thank you for sharing that with me. I'm glad you're feeling better now that you're
92
+ here. Sometimes waiting can cause anxiety or stress, so it's important to find
93
+ ways to manage these feelings. If you'd like, we can discuss some coping strategies
94
+ for dealing with stress while waiting or in similar situations.
95
+ - It seems like you've been having ups and downs with your learning experience.
96
+ It's great that you sometimes feel accomplished when you finally understand something.
97
+ When you're feeling upset or frustrated, it might be helpful to take a break and
98
+ come back to it later, as you've mentioned. This can give you a fresh perspective
99
+ and may help reduce stress. It's also important to remember that learning is a
100
+ process, and it's okay to experience difficulties along the way.
101
+ - source_sentence: I remember generating alternative predictions. We said that I might
102
+ get nervous but if I stay focused on what I want to do, I can probably do it.
103
+ It's just hard to believe when I'm in this situation.
104
+ sentences:
105
+ - It seems like you've been dealing with unexpected expenses and some impulsive
106
+ spending habits. By planning for a stricter budget and understanding the difference
107
+ between essential and non-essential expenses, you can start to develop healthy
108
+ financial habits. Have you considered creating a budget plan and setting aside
109
+ a specific amount for unexpected expenses to help manage your spending?
110
+ - It's normal to feel nervous in difficult situations. As a counsellor, I suggest
111
+ continuing to test these alternative predictions directly in real-life experiences.
112
+ As you encounter situations where you manage to handle things even while nervous,
113
+ your belief in your abilities and the predictions will strengthen. Remember to
114
+ stay focused and trust yourself.
115
+ - If you are a counsellor, please answer the questions based on the description
116
+ of the patient.
117
+ ---
118
+
119
+ # SentenceTransformer based on distilbert/distilroberta-base
120
+
121
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
122
+
123
+ ## Model Details
124
+
125
+ ### Model Description
126
+ - **Model Type:** Sentence Transformer
127
+ - **Base model:** [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base) <!-- at revision fb53ab8802853c8e4fbdbcd0529f21fc6f459b2b -->
128
+ - **Maximum Sequence Length:** 512 tokens
129
+ - **Output Dimensionality:** 768 tokens
130
+ - **Similarity Function:** Cosine Similarity
131
+ <!-- - **Training Dataset:** Unknown -->
132
+ <!-- - **Language:** Unknown -->
133
+ <!-- - **License:** Unknown -->
134
+
135
+ ### Model Sources
136
+
137
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
138
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
139
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
140
+
141
+ ### Full Model Architecture
142
+
143
+ ```
144
+ SentenceTransformer(
145
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
146
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
147
+ )
148
+ ```
149
+
150
+ ## Usage
151
+
152
+ ### Direct Usage (Sentence Transformers)
153
+
154
+ First install the Sentence Transformers library:
155
+
156
+ ```bash
157
+ pip install -U sentence-transformers
158
+ ```
159
+
160
+ Then you can load this model and run inference.
161
+ ```python
162
+ from sentence_transformers import SentenceTransformer
163
+
164
+ # Download from the 🤗 Hub
165
+ model = SentenceTransformer("iqrakiran/1my-sentence-transformer-model")
166
+ # Run inference
167
+ sentences = [
168
+ "I remember generating alternative predictions. We said that I might get nervous but if I stay focused on what I want to do, I can probably do it. It's just hard to believe when I'm in this situation.",
169
+ "It's normal to feel nervous in difficult situations. As a counsellor, I suggest continuing to test these alternative predictions directly in real-life experiences. As you encounter situations where you manage to handle things even while nervous, your belief in your abilities and the predictions will strengthen. Remember to stay focused and trust yourself.",
170
+ 'If you are a counsellor, please answer the questions based on the description of the patient.',
171
+ ]
172
+ embeddings = model.encode(sentences)
173
+ print(embeddings.shape)
174
+ # [3, 768]
175
+
176
+ # Get the similarity scores for the embeddings
177
+ similarities = model.similarity(embeddings, embeddings)
178
+ print(similarities.shape)
179
+ # [3, 3]
180
+ ```
181
+
182
+ <!--
183
+ ### Direct Usage (Transformers)
184
+
185
+ <details><summary>Click to see the direct usage in Transformers</summary>
186
+
187
+ </details>
188
+ -->
189
+
190
+ <!--
191
+ ### Downstream Usage (Sentence Transformers)
192
+
193
+ You can finetune this model on your own dataset.
194
+
195
+ <details><summary>Click to expand</summary>
196
+
197
+ </details>
198
+ -->
199
+
200
+ <!--
201
+ ### Out-of-Scope Use
202
+
203
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
204
+ -->
205
+
206
+ <!--
207
+ ## Bias, Risks and Limitations
208
+
209
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
210
+ -->
211
+
212
+ <!--
213
+ ### Recommendations
214
+
215
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
216
+ -->
217
+
218
+ ## Training Details
219
+
220
+ ### Training Dataset
221
+
222
+ #### Unnamed Dataset
223
+
224
+
225
+ * Size: 4,093 training samples
226
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
227
+ * Approximate statistics based on the first 1000 samples:
228
+ | | sentence_0 | sentence_1 | sentence_2 |
229
+ |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
230
+ | type | string | string | string |
231
+ | details | <ul><li>min: 12 tokens</li><li>mean: 56.8 tokens</li><li>max: 178 tokens</li></ul> | <ul><li>min: 27 tokens</li><li>mean: 76.32 tokens</li><li>max: 170 tokens</li></ul> | <ul><li>min: 22 tokens</li><li>mean: 22.0 tokens</li><li>max: 22 tokens</li></ul> |
232
+ * Samples:
233
+ | sentence_0 | sentence_1 | sentence_2 |
234
+ |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------|
235
+ | <code>I have been practicing muscle relaxation techniques to calm myself, but I still experience a lot of anxiety. When I'm relaxed and thinking about something calm, it feels nice. However, when I start to think about a negative situation, like people laughing at me, I quickly lose that feeling of relaxation.</code> | <code>You're on the right track with practicing muscle relaxation, but managing anxiety goes beyond the physical aspects. In social situations, it's important to shift our focus away from anxiety-provoking thoughts and towards the actions we can take in those situations. Remember, relaxation techniques are just one part of managing anxiety, so try to actively redirect your attention towards positive or neutral thoughts when negative scenarios come up.</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
236
+ | <code>As the patient, I feel disappointed and I'm really pissed about my current situation.</code> | <code>As your counsellor, I hear that you are experiencing strong emotions at the moment. It's important for us to identify the specific circumstances that have led to these feelings, so we can better address and challenge the irrational beliefs that might be contributing to your distress. Can you tell me more about the situation that has been bothering you?</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
237
+ | <code>I've been struggling with my relationship with my mom, and it has affected the way I interact with my kids. I vowed not to treat them the way she treated me, but I didn't realize how much it has affected the way I treat myself. I didn't deliberately want to be this way, but I ended up with something like an alter-ego.</code> | <code>It's important to recognize that your relationship with yourself is what we're focusing on right now, even though it was influenced by your relationship with your mom. It's not uncommon for people to unintentionally develop certain traits or behaviors due to past experiences. Acknowledge what you've discovered about yourself and try to gain a better understanding of the factors that contributed to it. Also, reflect on how to improve your relationship with yourself going forward. Have you considered any specific strategies or steps to improve your self-treatment?</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
238
+ * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
239
+ ```json
240
+ {
241
+ "scale": 20.0,
242
+ "similarity_fct": "cos_sim"
243
+ }
244
+ ```
245
+
246
+ ### Training Hyperparameters
247
+ #### Non-Default Hyperparameters
248
+
249
+ - `per_device_train_batch_size`: 16
250
+ - `per_device_eval_batch_size`: 16
251
+ - `num_train_epochs`: 10
252
+ - `multi_dataset_batch_sampler`: round_robin
253
+
254
+ #### All Hyperparameters
255
+ <details><summary>Click to expand</summary>
256
+
257
+ - `overwrite_output_dir`: False
258
+ - `do_predict`: False
259
+ - `eval_strategy`: no
260
+ - `prediction_loss_only`: True
261
+ - `per_device_train_batch_size`: 16
262
+ - `per_device_eval_batch_size`: 16
263
+ - `per_gpu_train_batch_size`: None
264
+ - `per_gpu_eval_batch_size`: None
265
+ - `gradient_accumulation_steps`: 1
266
+ - `eval_accumulation_steps`: None
267
+ - `torch_empty_cache_steps`: None
268
+ - `learning_rate`: 5e-05
269
+ - `weight_decay`: 0.0
270
+ - `adam_beta1`: 0.9
271
+ - `adam_beta2`: 0.999
272
+ - `adam_epsilon`: 1e-08
273
+ - `max_grad_norm`: 1
274
+ - `num_train_epochs`: 10
275
+ - `max_steps`: -1
276
+ - `lr_scheduler_type`: linear
277
+ - `lr_scheduler_kwargs`: {}
278
+ - `warmup_ratio`: 0.0
279
+ - `warmup_steps`: 0
280
+ - `log_level`: passive
281
+ - `log_level_replica`: warning
282
+ - `log_on_each_node`: True
283
+ - `logging_nan_inf_filter`: True
284
+ - `save_safetensors`: True
285
+ - `save_on_each_node`: False
286
+ - `save_only_model`: False
287
+ - `restore_callback_states_from_checkpoint`: False
288
+ - `no_cuda`: False
289
+ - `use_cpu`: False
290
+ - `use_mps_device`: False
291
+ - `seed`: 42
292
+ - `data_seed`: None
293
+ - `jit_mode_eval`: False
294
+ - `use_ipex`: False
295
+ - `bf16`: False
296
+ - `fp16`: False
297
+ - `fp16_opt_level`: O1
298
+ - `half_precision_backend`: auto
299
+ - `bf16_full_eval`: False
300
+ - `fp16_full_eval`: False
301
+ - `tf32`: None
302
+ - `local_rank`: 0
303
+ - `ddp_backend`: None
304
+ - `tpu_num_cores`: None
305
+ - `tpu_metrics_debug`: False
306
+ - `debug`: []
307
+ - `dataloader_drop_last`: False
308
+ - `dataloader_num_workers`: 0
309
+ - `dataloader_prefetch_factor`: None
310
+ - `past_index`: -1
311
+ - `disable_tqdm`: False
312
+ - `remove_unused_columns`: True
313
+ - `label_names`: None
314
+ - `load_best_model_at_end`: False
315
+ - `ignore_data_skip`: False
316
+ - `fsdp`: []
317
+ - `fsdp_min_num_params`: 0
318
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
319
+ - `fsdp_transformer_layer_cls_to_wrap`: None
320
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
321
+ - `deepspeed`: None
322
+ - `label_smoothing_factor`: 0.0
323
+ - `optim`: adamw_torch
324
+ - `optim_args`: None
325
+ - `adafactor`: False
326
+ - `group_by_length`: False
327
+ - `length_column_name`: length
328
+ - `ddp_find_unused_parameters`: None
329
+ - `ddp_bucket_cap_mb`: None
330
+ - `ddp_broadcast_buffers`: False
331
+ - `dataloader_pin_memory`: True
332
+ - `dataloader_persistent_workers`: False
333
+ - `skip_memory_metrics`: True
334
+ - `use_legacy_prediction_loop`: False
335
+ - `push_to_hub`: False
336
+ - `resume_from_checkpoint`: None
337
+ - `hub_model_id`: None
338
+ - `hub_strategy`: every_save
339
+ - `hub_private_repo`: False
340
+ - `hub_always_push`: False
341
+ - `gradient_checkpointing`: False
342
+ - `gradient_checkpointing_kwargs`: None
343
+ - `include_inputs_for_metrics`: False
344
+ - `eval_do_concat_batches`: True
345
+ - `fp16_backend`: auto
346
+ - `push_to_hub_model_id`: None
347
+ - `push_to_hub_organization`: None
348
+ - `mp_parameters`:
349
+ - `auto_find_batch_size`: False
350
+ - `full_determinism`: False
351
+ - `torchdynamo`: None
352
+ - `ray_scope`: last
353
+ - `ddp_timeout`: 1800
354
+ - `torch_compile`: False
355
+ - `torch_compile_backend`: None
356
+ - `torch_compile_mode`: None
357
+ - `dispatch_batches`: None
358
+ - `split_batches`: None
359
+ - `include_tokens_per_second`: False
360
+ - `include_num_input_tokens_seen`: False
361
+ - `neftune_noise_alpha`: None
362
+ - `optim_target_modules`: None
363
+ - `batch_eval_metrics`: False
364
+ - `eval_on_start`: False
365
+ - `eval_use_gather_object`: False
366
+ - `batch_sampler`: batch_sampler
367
+ - `multi_dataset_batch_sampler`: round_robin
368
+
369
+ </details>
370
+
371
+ ### Training Logs
372
+ | Epoch | Step | Training Loss |
373
+ |:------:|:----:|:-------------:|
374
+ | 1.9531 | 500 | 0.2321 |
375
+ | 3.9062 | 1000 | 0.0 |
376
+ | 5.8594 | 1500 | 0.0001 |
377
+ | 7.8125 | 2000 | 0.0 |
378
+ | 9.7656 | 2500 | 0.0 |
379
+ | 1.9531 | 500 | 0.4353 |
380
+ | 3.9062 | 1000 | 0.0119 |
381
+ | 5.8594 | 1500 | 0.0022 |
382
+ | 7.8125 | 2000 | 0.0009 |
383
+ | 9.7656 | 2500 | 0.0007 |
384
+
385
+
386
+ ### Framework Versions
387
+ - Python: 3.10.12
388
+ - Sentence Transformers: 3.0.1
389
+ - Transformers: 4.44.2
390
+ - PyTorch: 2.4.0+cu121
391
+ - Accelerate: 0.33.0
392
+ - Datasets: 2.21.0
393
+ - Tokenizers: 0.19.1
394
+
395
+ ## Citation
396
+
397
+ ### BibTeX
398
+
399
+ #### Sentence Transformers
400
+ ```bibtex
401
+ @inproceedings{reimers-2019-sentence-bert,
402
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
403
+ author = "Reimers, Nils and Gurevych, Iryna",
404
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
405
+ month = "11",
406
+ year = "2019",
407
+ publisher = "Association for Computational Linguistics",
408
+ url = "https://arxiv.org/abs/1908.10084",
409
+ }
410
+ ```
411
+
412
+ #### MultipleNegativesRankingLoss
413
+ ```bibtex
414
+ @misc{henderson2017efficient,
415
+ title={Efficient Natural Language Response Suggestion for Smart Reply},
416
+ author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
417
+ year={2017},
418
+ eprint={1705.00652},
419
+ archivePrefix={arXiv},
420
+ primaryClass={cs.CL}
421
+ }
422
+ ```
423
+
424
+ <!--
425
+ ## Glossary
426
+
427
+ *Clearly define terms in order to be accessible across audiences.*
428
+ -->
429
+
430
+ <!--
431
+ ## Model Card Authors
432
+
433
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
434
+ -->
435
+
436
+ <!--
437
+ ## Model Card Contact
438
+
439
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
440
+ -->
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilroberta-base",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 6,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.44.2",
24
+ "type_vocab_size": 1,
25
+ "use_cache": true,
26
+ "vocab_size": 50265
27
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.44.2",
5
+ "pytorch": "2.4.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf472ad31c5a34e5fc835d368cb9051a1d07f6d548a4846d327f5d46c6631f2f
3
+ size 328485128
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "mask_token": "<mask>",
51
+ "model_max_length": 512,
52
+ "pad_token": "<pad>",
53
+ "sep_token": "</s>",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
+ "unk_token": "<unk>"
57
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff