mohityadav commited on
Commit
a184792
1 Parent(s): d393d46

mohityadav/mental-health-advisor-gpt

Browse files
README.md ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ base_model: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
7
+ model-index:
8
+ - name: working
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # working
16
+
17
+ This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.2984
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0002
39
+ - train_batch_size: 4
40
+ - eval_batch_size: 4
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 16
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 2
47
+ - num_epochs: 10
48
+ - mixed_precision_training: Native AMP
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:------:|:----:|:---------------:|
54
+ | 3.1139 | 0.9412 | 12 | 2.3084 |
55
+ | 1.9191 | 1.9608 | 25 | 1.7055 |
56
+ | 1.6083 | 2.9804 | 38 | 1.5741 |
57
+ | 1.496 | 4.0 | 51 | 1.4695 |
58
+ | 1.5309 | 4.9412 | 63 | 1.4038 |
59
+ | 1.3627 | 5.9608 | 76 | 1.3573 |
60
+ | 1.3111 | 6.9804 | 89 | 1.3281 |
61
+ | 1.2995 | 8.0 | 102 | 1.3092 |
62
+ | 1.3866 | 8.9412 | 114 | 1.2998 |
63
+ | 1.18 | 9.4118 | 120 | 1.2984 |
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - PEFT 0.11.1
69
+ - Transformers 4.41.2
70
+ - Pytorch 2.1.2
71
+ - Datasets 2.19.2
72
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "q_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM",
26
+ "use_dora": false,
27
+ "use_rslora": false
28
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cf94b36f046daba8d3cb473ba0348657a9ab8c56dbc2fb787002a64c357e7fa
3
+ size 8401408
runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385345.adc95cf38b20.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bca8edc01aa31a9e70f9815dbca96d3ac87c4247cc92a4349a8cb745bdc575e3
3
+ size 5481
runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385423.adc95cf38b20.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b0060337b7a08396a73032f8dc408d21d457d6d92945573baaee52a0267a943
3
+ size 5481
runs/Jun26_07-07-26_adc95cf38b20/events.out.tfevents.1719385648.adc95cf38b20.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcba8cc9427be248c5ef089831420e28a59559d37e4d7b1fc100ffbe95dd1d22
3
+ size 16267
runs/Jun26_07-16-15_adc95cf38b20/events.out.tfevents.1719386177.adc95cf38b20.34.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71bffc05745c6f8799b41a6de611947dd2d1daf0eef79e856f120996c9f55035
3
+ size 5481
runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719386347.adc95cf38b20.34.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e7083d5109e346d61a1145b6a1fdbd73bcf0549de48bd847c6788e0072883c2
3
+ size 10559
runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719391099.adc95cf38b20.34.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1664a1d1de74046bcd26b750d0bb5b1b704fb72256c794a139ae46f340dc4c85
3
+ size 354
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be11ac523aaa8727f2185f783bf03400ad8ca8ff6208855f6a032bd7b8a919b3
3
+ size 5112
wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
10
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
11
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
12
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
13
+ 2024-06-26 07:16:28,967 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
14
+ Traceback (most recent call last):
15
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
16
+ exec(code_obj, self.user_global_ns, self.user_ns)
17
+ File "/tmp/ipykernel_34/2014566126.py", line 10, in <module>
18
+ trainer.train()
19
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
20
+ return inner_training_loop(
21
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
22
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
23
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
24
+ return self.call_event("on_train_begin", args, state, control)
25
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
26
+ result = getattr(callback, event)(
27
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
28
+ self.setup(args, state, model, **kwargs)
29
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
30
+ self._wandb.init(
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
32
+ wandb._sentry.reraise(e)
33
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
34
+ raise exc.with_traceback(sys.exc_info()[2])
35
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
36
+ wi.setup(kwargs)
37
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
38
+ wandb_login._login(
39
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
40
+ wlogin.prompt_api_key()
41
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
42
+ key, status = self._prompt_api_key()
43
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
44
+ key = apikey.prompt_api_key(
45
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
46
+ key = input_callback(api_ask).strip()
47
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
48
+ value = prompt_func(prompt)
49
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
50
+ raise Abort() from None
51
+ click.exceptions.Abort
52
+ 2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
53
+ Traceback (most recent call last):
54
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
55
+ exec(code_obj, self.user_global_ns, self.user_ns)
56
+ File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
57
+ trainer.train()
58
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
59
+ return inner_training_loop(
60
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
61
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
62
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
63
+ return self.call_event("on_train_begin", args, state, control)
64
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
65
+ result = getattr(callback, event)(
66
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
67
+ self.setup(args, state, model, **kwargs)
68
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
69
+ self._wandb.init(
70
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
71
+ wandb._sentry.reraise(e)
72
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
73
+ raise exc.with_traceback(sys.exc_info()[2])
74
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
75
+ wi.setup(kwargs)
76
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
77
+ wandb_login._login(
78
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
79
+ wlogin.prompt_api_key()
80
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
81
+ key, status = self._prompt_api_key()
82
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
83
+ key = apikey.prompt_api_key(
84
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
85
+ key = input_callback(api_ask).strip()
86
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
87
+ value = prompt_func(prompt)
88
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
89
+ raise Abort() from None
90
+ click.exceptions.Abort
91
+ 2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
92
+ Traceback (most recent call last):
93
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
94
+ exec(code_obj, self.user_global_ns, self.user_ns)
95
+ File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
96
+ trainer.train()
97
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
98
+ return inner_training_loop(
99
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
100
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
101
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
102
+ return self.call_event("on_train_begin", args, state, control)
103
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
104
+ result = getattr(callback, event)(
105
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
106
+ self.setup(args, state, model, **kwargs)
107
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
108
+ self._wandb.init(
109
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
110
+ wandb._sentry.reraise(e)
111
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
112
+ raise exc.with_traceback(sys.exc_info()[2])
113
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
114
+ wi.setup(kwargs)
115
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
116
+ wandb_login._login(
117
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
118
+ wlogin.prompt_api_key()
119
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
120
+ key, status = self._prompt_api_key()
121
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
122
+ key = apikey.prompt_api_key(
123
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
124
+ key = input_callback(api_ask).strip()
125
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
126
+ value = prompt_func(prompt)
127
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
128
+ raise Abort() from None
129
+ click.exceptions.Abort
130
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug.log
131
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log
132
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78720aaa5b40>
133
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
134
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
135
+ config: {}
136
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():610] starting backend
137
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
138
+ 2024-06-26 07:16:28,971 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
139
+ 2024-06-26 07:16:28,977 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
140
+ 2024-06-26 07:16:28,989 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
141
+ 2024-06-26 07:16:29,303 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
142
+ 2024-06-26 07:16:29,306 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
143
+ 2024-06-26 07:16:29,429 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
144
+ 2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
145
+
146
+ 2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
147
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
148
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
149
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
150
+ 2024-06-26 07:16:45,663 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
151
+ 2024-06-26 07:16:45,672 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
152
+ 2024-06-26 07:16:45,678 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-16-15_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
153
+ 2024-06-26 07:18:26,851 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
154
+ 2024-06-26 07:18:26,851 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
155
+ 2024-06-26 07:18:50,892 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
156
+ 2024-06-26 07:18:50,915 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
157
+ 2024-06-26 07:18:50,915 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
158
+ 2024-06-26 07:18:51,909 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
159
+ 2024-06-26 07:18:51,913 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
160
+ 2024-06-26 07:18:51,913 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
161
+ 2024-06-26 07:18:52,853 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
162
+ 2024-06-26 07:18:52,855 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
163
+ 2024-06-26 07:18:52,855 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
164
+ 2024-06-26 07:18:53,667 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
165
+ 2024-06-26 07:18:53,670 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
166
+ 2024-06-26 07:18:53,670 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
167
+ 2024-06-26 07:18:54,691 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
168
+ 2024-06-26 07:18:54,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
169
+ 2024-06-26 07:18:54,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
170
+ 2024-06-26 07:18:55,617 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
171
+ 2024-06-26 07:18:55,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
172
+ 2024-06-26 07:18:55,682 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
173
+ 2024-06-26 07:18:56,512 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
174
+ 2024-06-26 07:18:56,513 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
175
+ 2024-06-26 07:18:56,513 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
176
+ 2024-06-26 07:18:57,999 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
177
+ 2024-06-26 07:18:58,001 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
178
+ 2024-06-26 07:18:58,001 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
179
+ 2024-06-26 07:18:58,997 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
180
+ 2024-06-26 07:18:59,009 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
181
+ 2024-06-26 07:18:59,009 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
182
+ 2024-06-26 07:19:00,090 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
183
+ 2024-06-26 07:19:00,092 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
184
+ 2024-06-26 07:19:00,092 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
185
+ 2024-06-26 07:19:01,722 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
186
+ 2024-06-26 07:19:01,723 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
187
+ 2024-06-26 07:19:01,723 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
188
+ 2024-06-26 07:19:03,098 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
189
+ 2024-06-26 07:19:03,117 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
190
+ 2024-06-26 07:19:03,117 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
191
+ 2024-06-26 07:19:04,117 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
192
+ 2024-06-26 07:19:04,212 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
193
+ 2024-06-26 07:19:04,212 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
194
+ 2024-06-26 07:19:05,065 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
195
+ 2024-06-26 07:19:05,097 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
196
+ 2024-06-26 07:19:05,097 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
197
+ 2024-06-26 07:19:06,209 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
198
+ 2024-06-26 07:19:06,220 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
199
+ 2024-06-26 07:19:06,220 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
200
+ 2024-06-26 07:19:07,300 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
201
+ 2024-06-26 07:19:07,961 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-19-05_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
202
+ 2024-06-26 08:35:48,386 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
203
+ 2024-06-26 08:35:48,387 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
204
+ 2024-06-26 08:36:11,677 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
205
+ 2024-06-26 08:38:19,575 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
206
+ 2024-06-26 08:38:19,575 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
207
+ 2024-06-26 08:39:55,100 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
208
+ 2024-06-26 08:39:55,103 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
209
+ 2024-06-26 08:39:55,103 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
210
+ 2024-06-26 08:40:32,017 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
211
+ 2024-06-26 08:43:02,593 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
212
+ 2024-06-26 08:43:02,594 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
213
+ 2024-06-26 08:45:10,959 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
214
+ 2024-06-26 08:45:10,963 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
215
+ 2024-06-26 08:45:10,963 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
216
+ 2024-06-26 08:45:13,384 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
217
+ 2024-06-26 08:47:43,994 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
218
+ 2024-06-26 08:47:43,994 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
219
+ 2024-06-26 08:49:58,760 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
220
+ 2024-06-26 08:53:33,771 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
221
+ 2024-06-26 08:53:33,771 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
222
+ 2024-06-26 08:54:32,055 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
223
+ 2024-06-26 08:54:32,059 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
224
+ 2024-06-26 08:54:32,059 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
225
+ 2024-06-26 08:54:32,754 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
226
+ 2024-06-26 08:58:07,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
227
+ 2024-06-26 08:58:07,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
228
+ 2024-06-26 09:12:46,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
229
+ 2024-06-26 09:12:46,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
230
+ 2024-06-26 09:12:46,669 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
231
+ 2024-06-26 09:16:22,037 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
232
+ 2024-06-26 09:16:22,038 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
233
+ 2024-06-26 09:16:22,038 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
234
+ 2024-06-26 09:16:35,228 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
235
+ 2024-06-26 09:16:36,438 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
236
+ 2024-06-26 09:16:36,438 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
237
+ 2024-06-26 09:18:25,016 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
238
+ 2024-06-26 09:18:26,213 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
239
+ 2024-06-26 09:18:26,214 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
240
+ 2024-06-26 09:21:21,819 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
241
+ 2024-06-26 09:21:21,842 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
242
+ 2024-06-26 09:21:21,842 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
243
+ 2024-06-26 09:21:36,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
244
+ 2024-06-26 09:21:36,633 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
245
+ 2024-06-26 09:21:36,634 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
246
+ 2024-06-26 09:21:37,142 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240626_071628-smnm2aje/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240626_071628-smnm2aje/files/config.yaml ADDED
@@ -0,0 +1,724 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.17.0
8
+ framework: huggingface
9
+ huggingface_version: 4.41.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1719386188
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 99
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 98
42
+ - 99
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 13
47
+ - 23
48
+ - 62
49
+ - 66
50
+ 4: 3.10.13
51
+ 5: 0.17.0
52
+ 6: 4.41.2
53
+ 8:
54
+ - 1
55
+ - 2
56
+ - 5
57
+ 9:
58
+ 1: transformers_trainer
59
+ 13: linux-x86_64
60
+ m:
61
+ - 1: train/global_step
62
+ 6:
63
+ - 3
64
+ - 1: train/loss
65
+ 5: 1
66
+ 6:
67
+ - 1
68
+ - 1: train/grad_norm
69
+ 5: 1
70
+ 6:
71
+ - 1
72
+ - 1: train/learning_rate
73
+ 5: 1
74
+ 6:
75
+ - 1
76
+ - 1: train/epoch
77
+ 5: 1
78
+ 6:
79
+ - 1
80
+ - 1: eval/loss
81
+ 5: 1
82
+ 6:
83
+ - 1
84
+ - 1: eval/runtime
85
+ 5: 1
86
+ 6:
87
+ - 1
88
+ - 1: eval/samples_per_second
89
+ 5: 1
90
+ 6:
91
+ - 1
92
+ - 1: eval/steps_per_second
93
+ 5: 1
94
+ 6:
95
+ - 1
96
+ vocab_size:
97
+ desc: null
98
+ value: 32000
99
+ max_position_embeddings:
100
+ desc: null
101
+ value: 32768
102
+ hidden_size:
103
+ desc: null
104
+ value: 4096
105
+ intermediate_size:
106
+ desc: null
107
+ value: 14336
108
+ num_hidden_layers:
109
+ desc: null
110
+ value: 32
111
+ num_attention_heads:
112
+ desc: null
113
+ value: 32
114
+ sliding_window:
115
+ desc: null
116
+ value: null
117
+ num_key_value_heads:
118
+ desc: null
119
+ value: 8
120
+ hidden_act:
121
+ desc: null
122
+ value: silu
123
+ initializer_range:
124
+ desc: null
125
+ value: 0.02
126
+ rms_norm_eps:
127
+ desc: null
128
+ value: 1.0e-05
129
+ use_cache:
130
+ desc: null
131
+ value: false
132
+ rope_theta:
133
+ desc: null
134
+ value: 1000000.0
135
+ attention_dropout:
136
+ desc: null
137
+ value: 0.0
138
+ return_dict:
139
+ desc: null
140
+ value: true
141
+ output_hidden_states:
142
+ desc: null
143
+ value: false
144
+ output_attentions:
145
+ desc: null
146
+ value: false
147
+ torchscript:
148
+ desc: null
149
+ value: false
150
+ torch_dtype:
151
+ desc: null
152
+ value: bfloat16
153
+ use_bfloat16:
154
+ desc: null
155
+ value: false
156
+ tf_legacy_loss:
157
+ desc: null
158
+ value: false
159
+ pruned_heads:
160
+ desc: null
161
+ value: {}
162
+ tie_word_embeddings:
163
+ desc: null
164
+ value: false
165
+ chunk_size_feed_forward:
166
+ desc: null
167
+ value: 0
168
+ is_encoder_decoder:
169
+ desc: null
170
+ value: false
171
+ is_decoder:
172
+ desc: null
173
+ value: false
174
+ cross_attention_hidden_size:
175
+ desc: null
176
+ value: null
177
+ add_cross_attention:
178
+ desc: null
179
+ value: false
180
+ tie_encoder_decoder:
181
+ desc: null
182
+ value: false
183
+ max_length:
184
+ desc: null
185
+ value: 20
186
+ min_length:
187
+ desc: null
188
+ value: 0
189
+ do_sample:
190
+ desc: null
191
+ value: false
192
+ early_stopping:
193
+ desc: null
194
+ value: false
195
+ num_beams:
196
+ desc: null
197
+ value: 1
198
+ num_beam_groups:
199
+ desc: null
200
+ value: 1
201
+ diversity_penalty:
202
+ desc: null
203
+ value: 0.0
204
+ temperature:
205
+ desc: null
206
+ value: 1.0
207
+ top_k:
208
+ desc: null
209
+ value: 50
210
+ top_p:
211
+ desc: null
212
+ value: 1.0
213
+ typical_p:
214
+ desc: null
215
+ value: 1.0
216
+ repetition_penalty:
217
+ desc: null
218
+ value: 1.0
219
+ length_penalty:
220
+ desc: null
221
+ value: 1.0
222
+ no_repeat_ngram_size:
223
+ desc: null
224
+ value: 0
225
+ encoder_no_repeat_ngram_size:
226
+ desc: null
227
+ value: 0
228
+ bad_words_ids:
229
+ desc: null
230
+ value: null
231
+ num_return_sequences:
232
+ desc: null
233
+ value: 1
234
+ output_scores:
235
+ desc: null
236
+ value: false
237
+ return_dict_in_generate:
238
+ desc: null
239
+ value: false
240
+ forced_bos_token_id:
241
+ desc: null
242
+ value: null
243
+ forced_eos_token_id:
244
+ desc: null
245
+ value: null
246
+ remove_invalid_values:
247
+ desc: null
248
+ value: false
249
+ exponential_decay_length_penalty:
250
+ desc: null
251
+ value: null
252
+ suppress_tokens:
253
+ desc: null
254
+ value: null
255
+ begin_suppress_tokens:
256
+ desc: null
257
+ value: null
258
+ architectures:
259
+ desc: null
260
+ value:
261
+ - MistralForCausalLM
262
+ finetuning_task:
263
+ desc: null
264
+ value: null
265
+ id2label:
266
+ desc: null
267
+ value:
268
+ '0': LABEL_0
269
+ '1': LABEL_1
270
+ label2id:
271
+ desc: null
272
+ value:
273
+ LABEL_0: 0
274
+ LABEL_1: 1
275
+ tokenizer_class:
276
+ desc: null
277
+ value: null
278
+ prefix:
279
+ desc: null
280
+ value: null
281
+ bos_token_id:
282
+ desc: null
283
+ value: 1
284
+ pad_token_id:
285
+ desc: null
286
+ value: 0
287
+ eos_token_id:
288
+ desc: null
289
+ value: 2
290
+ sep_token_id:
291
+ desc: null
292
+ value: null
293
+ decoder_start_token_id:
294
+ desc: null
295
+ value: null
296
+ task_specific_params:
297
+ desc: null
298
+ value: null
299
+ problem_type:
300
+ desc: null
301
+ value: null
302
+ _name_or_path:
303
+ desc: null
304
+ value: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
305
+ transformers_version:
306
+ desc: null
307
+ value: 4.41.2
308
+ model_type:
309
+ desc: null
310
+ value: mistral
311
+ pretraining_tp:
312
+ desc: null
313
+ value: 1
314
+ quantization_config:
315
+ desc: null
316
+ value:
317
+ quant_method: QuantizationMethod.GPTQ
318
+ bits: 4
319
+ tokenizer: null
320
+ dataset: null
321
+ group_size: 128
322
+ damp_percent: 0.1
323
+ desc_act: true
324
+ sym: true
325
+ true_sequential: true
326
+ use_cuda_fp16: false
327
+ model_seqlen: null
328
+ block_name_to_quantize: null
329
+ module_name_preceding_first_block: null
330
+ batch_size: 1
331
+ pad_token_id: null
332
+ use_exllama: true
333
+ max_input_length: null
334
+ exllama_config:
335
+ version: ExllamaVersion.ONE
336
+ cache_block_outputs: true
337
+ modules_in_block_to_quantize: null
338
+ output_dir:
339
+ desc: null
340
+ value: /kaggle/working/
341
+ overwrite_output_dir:
342
+ desc: null
343
+ value: false
344
+ do_train:
345
+ desc: null
346
+ value: false
347
+ do_eval:
348
+ desc: null
349
+ value: true
350
+ do_predict:
351
+ desc: null
352
+ value: false
353
+ eval_strategy:
354
+ desc: null
355
+ value: epoch
356
+ prediction_loss_only:
357
+ desc: null
358
+ value: false
359
+ per_device_train_batch_size:
360
+ desc: null
361
+ value: 4
362
+ per_device_eval_batch_size:
363
+ desc: null
364
+ value: 4
365
+ per_gpu_train_batch_size:
366
+ desc: null
367
+ value: null
368
+ per_gpu_eval_batch_size:
369
+ desc: null
370
+ value: null
371
+ gradient_accumulation_steps:
372
+ desc: null
373
+ value: 4
374
+ eval_accumulation_steps:
375
+ desc: null
376
+ value: null
377
+ eval_delay:
378
+ desc: null
379
+ value: 0
380
+ learning_rate:
381
+ desc: null
382
+ value: 0.0002
383
+ weight_decay:
384
+ desc: null
385
+ value: 0.01
386
+ adam_beta1:
387
+ desc: null
388
+ value: 0.9
389
+ adam_beta2:
390
+ desc: null
391
+ value: 0.999
392
+ adam_epsilon:
393
+ desc: null
394
+ value: 1.0e-08
395
+ max_grad_norm:
396
+ desc: null
397
+ value: 1.0
398
+ num_train_epochs:
399
+ desc: null
400
+ value: 10
401
+ max_steps:
402
+ desc: null
403
+ value: -1
404
+ lr_scheduler_type:
405
+ desc: null
406
+ value: linear
407
+ lr_scheduler_kwargs:
408
+ desc: null
409
+ value: {}
410
+ warmup_ratio:
411
+ desc: null
412
+ value: 0.0
413
+ warmup_steps:
414
+ desc: null
415
+ value: 2
416
+ log_level:
417
+ desc: null
418
+ value: passive
419
+ log_level_replica:
420
+ desc: null
421
+ value: warning
422
+ log_on_each_node:
423
+ desc: null
424
+ value: true
425
+ logging_dir:
426
+ desc: null
427
+ value: /kaggle/working/runs/Jun26_07-19-05_adc95cf38b20
428
+ logging_strategy:
429
+ desc: null
430
+ value: epoch
431
+ logging_first_step:
432
+ desc: null
433
+ value: false
434
+ logging_steps:
435
+ desc: null
436
+ value: 500
437
+ logging_nan_inf_filter:
438
+ desc: null
439
+ value: true
440
+ save_strategy:
441
+ desc: null
442
+ value: epoch
443
+ save_steps:
444
+ desc: null
445
+ value: 500
446
+ save_total_limit:
447
+ desc: null
448
+ value: null
449
+ save_safetensors:
450
+ desc: null
451
+ value: true
452
+ save_on_each_node:
453
+ desc: null
454
+ value: false
455
+ save_only_model:
456
+ desc: null
457
+ value: false
458
+ restore_callback_states_from_checkpoint:
459
+ desc: null
460
+ value: false
461
+ no_cuda:
462
+ desc: null
463
+ value: false
464
+ use_cpu:
465
+ desc: null
466
+ value: false
467
+ use_mps_device:
468
+ desc: null
469
+ value: false
470
+ seed:
471
+ desc: null
472
+ value: 42
473
+ data_seed:
474
+ desc: null
475
+ value: null
476
+ jit_mode_eval:
477
+ desc: null
478
+ value: false
479
+ use_ipex:
480
+ desc: null
481
+ value: false
482
+ bf16:
483
+ desc: null
484
+ value: false
485
+ fp16:
486
+ desc: null
487
+ value: true
488
+ fp16_opt_level:
489
+ desc: null
490
+ value: O1
491
+ half_precision_backend:
492
+ desc: null
493
+ value: auto
494
+ bf16_full_eval:
495
+ desc: null
496
+ value: false
497
+ fp16_full_eval:
498
+ desc: null
499
+ value: false
500
+ tf32:
501
+ desc: null
502
+ value: null
503
+ local_rank:
504
+ desc: null
505
+ value: 0
506
+ ddp_backend:
507
+ desc: null
508
+ value: null
509
+ tpu_num_cores:
510
+ desc: null
511
+ value: null
512
+ tpu_metrics_debug:
513
+ desc: null
514
+ value: false
515
+ debug:
516
+ desc: null
517
+ value: []
518
+ dataloader_drop_last:
519
+ desc: null
520
+ value: false
521
+ eval_steps:
522
+ desc: null
523
+ value: null
524
+ dataloader_num_workers:
525
+ desc: null
526
+ value: 0
527
+ dataloader_prefetch_factor:
528
+ desc: null
529
+ value: null
530
+ past_index:
531
+ desc: null
532
+ value: -1
533
+ run_name:
534
+ desc: null
535
+ value: /kaggle/working/
536
+ disable_tqdm:
537
+ desc: null
538
+ value: false
539
+ remove_unused_columns:
540
+ desc: null
541
+ value: true
542
+ label_names:
543
+ desc: null
544
+ value: null
545
+ load_best_model_at_end:
546
+ desc: null
547
+ value: true
548
+ metric_for_best_model:
549
+ desc: null
550
+ value: loss
551
+ greater_is_better:
552
+ desc: null
553
+ value: false
554
+ ignore_data_skip:
555
+ desc: null
556
+ value: false
557
+ fsdp:
558
+ desc: null
559
+ value: []
560
+ fsdp_min_num_params:
561
+ desc: null
562
+ value: 0
563
+ fsdp_config:
564
+ desc: null
565
+ value:
566
+ min_num_params: 0
567
+ xla: false
568
+ xla_fsdp_v2: false
569
+ xla_fsdp_grad_ckpt: false
570
+ fsdp_transformer_layer_cls_to_wrap:
571
+ desc: null
572
+ value: null
573
+ accelerator_config:
574
+ desc: null
575
+ value:
576
+ split_batches: false
577
+ dispatch_batches: null
578
+ even_batches: true
579
+ use_seedable_sampler: true
580
+ non_blocking: false
581
+ gradient_accumulation_kwargs: null
582
+ deepspeed:
583
+ desc: null
584
+ value: null
585
+ label_smoothing_factor:
586
+ desc: null
587
+ value: 0.0
588
+ optim:
589
+ desc: null
590
+ value: paged_adamw_8bit
591
+ optim_args:
592
+ desc: null
593
+ value: null
594
+ adafactor:
595
+ desc: null
596
+ value: false
597
+ group_by_length:
598
+ desc: null
599
+ value: false
600
+ length_column_name:
601
+ desc: null
602
+ value: length
603
+ report_to:
604
+ desc: null
605
+ value:
606
+ - tensorboard
607
+ - wandb
608
+ ddp_find_unused_parameters:
609
+ desc: null
610
+ value: null
611
+ ddp_bucket_cap_mb:
612
+ desc: null
613
+ value: null
614
+ ddp_broadcast_buffers:
615
+ desc: null
616
+ value: null
617
+ dataloader_pin_memory:
618
+ desc: null
619
+ value: true
620
+ dataloader_persistent_workers:
621
+ desc: null
622
+ value: false
623
+ skip_memory_metrics:
624
+ desc: null
625
+ value: true
626
+ use_legacy_prediction_loop:
627
+ desc: null
628
+ value: false
629
+ push_to_hub:
630
+ desc: null
631
+ value: false
632
+ resume_from_checkpoint:
633
+ desc: null
634
+ value: null
635
+ hub_model_id:
636
+ desc: null
637
+ value: null
638
+ hub_strategy:
639
+ desc: null
640
+ value: every_save
641
+ hub_token:
642
+ desc: null
643
+ value: <HUB_TOKEN>
644
+ hub_private_repo:
645
+ desc: null
646
+ value: false
647
+ hub_always_push:
648
+ desc: null
649
+ value: false
650
+ gradient_checkpointing:
651
+ desc: null
652
+ value: false
653
+ gradient_checkpointing_kwargs:
654
+ desc: null
655
+ value: null
656
+ include_inputs_for_metrics:
657
+ desc: null
658
+ value: false
659
+ eval_do_concat_batches:
660
+ desc: null
661
+ value: true
662
+ fp16_backend:
663
+ desc: null
664
+ value: auto
665
+ evaluation_strategy:
666
+ desc: null
667
+ value: null
668
+ push_to_hub_model_id:
669
+ desc: null
670
+ value: null
671
+ push_to_hub_organization:
672
+ desc: null
673
+ value: null
674
+ push_to_hub_token:
675
+ desc: null
676
+ value: <PUSH_TO_HUB_TOKEN>
677
+ mp_parameters:
678
+ desc: null
679
+ value: ''
680
+ auto_find_batch_size:
681
+ desc: null
682
+ value: false
683
+ full_determinism:
684
+ desc: null
685
+ value: false
686
+ torchdynamo:
687
+ desc: null
688
+ value: null
689
+ ray_scope:
690
+ desc: null
691
+ value: last
692
+ ddp_timeout:
693
+ desc: null
694
+ value: 1800
695
+ torch_compile:
696
+ desc: null
697
+ value: false
698
+ torch_compile_backend:
699
+ desc: null
700
+ value: null
701
+ torch_compile_mode:
702
+ desc: null
703
+ value: null
704
+ dispatch_batches:
705
+ desc: null
706
+ value: null
707
+ split_batches:
708
+ desc: null
709
+ value: null
710
+ include_tokens_per_second:
711
+ desc: null
712
+ value: false
713
+ include_num_input_tokens_seen:
714
+ desc: null
715
+ value: false
716
+ neftune_noise_alpha:
717
+ desc: null
718
+ value: null
719
+ optim_target_modules:
720
+ desc: null
721
+ value: null
722
+ batch_eval_metrics:
723
+ desc: null
724
+ value: false
wandb/run-20240626_071628-smnm2aje/files/output.log ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
2
+ warnings.warn(
3
+ KeyboardInterrupt
4
+ <s>[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
5
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
6
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
7
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
8
+ please provide the valuable advice for this comment.
9
+ I'm going through some things with my feelings and myself. I barely sleep and I do nothing but think about how I'm worthless and how I shouldn't be here.
10
+ I've never tried or contemplated suicide. I've always wanted to fix my issues, but I never get around to it.
11
+ How can I change my feeling of being worthless to everyone?
12
+ [/INST]
13
+ If everyone thinks you're worthless, then maybe you need to find new people to hang out with.Seriously, the social context in which a person lives is a big influence in self-esteem.Otherwise, you can go round and round trying to understand why you're not worthless, then go back to the same crowd and be knocked down again.There are many inspirational messages you can find in social media. Maybe read some of the ones which state that no person is worthless, and that everyone has a good purpose to their life.Also, since our culture is so saturated with the belief that if someone doesn't feel good about themselves that this is somehow terrible.Bad feelings are part of living. They are the motivation to remove ourselves from situations and relationships which do us more harm than good.Bad feelings do feel terrible. Your feeling of worthlessness may be good in the sense of motivating you to find out that you are much better than your feelings today.
14
+ </s>
15
+ <s>[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
16
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
17
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
18
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
19
+ please provide the valuable advice for this comment.
20
+ I've never been able to talk with my parents. My parents are in their sixties while I am a teenager. I love both of them but not their personalities. I feel that they do not take me seriously whenever I talk about a serious event in my life. If my dad doesn’t believe me, then my mom goes along with my dad and acts like she doesn’t believe me either. I’m a pansexual, but I can’t trust my own parents.
21
+ I've fought depression and won; however, stress and anxiety are killing me. I feel that my friends don't listen to me. I know they have their own problems, which I do my best to help with. But they don't always try to help me with mine, when I really need them. I feel as if my childhood has been taken from me. I feel as if I have no one whom I can trust.
22
+ [/INST]
23
+ Life sounds extremely frustrating for you right now.As long as you've reached your conclusion that your parents do not take you seriously when you would like to speak with them on a serious topic, and that you feel unable to trust either of them, then work with your own information.Many people don't know how to listen and aren't trustworthy. As painful and aggravating as having parents who fit this description, you're better off to honor what you realize than to continue fighting what is not in your hands to change.Since you write about having a similar problem with friends, then maybe this comes about because you pick friends who are similarly unsatisfying to be with as are your parents.The first step out of this dilemma would be to accept how crummy you feel about the way your parents are toward you.If you're able to live with the sadness and resentment of having parents who don't want to know how to understand you, then you'll start to develop a new and more satisfying definition of what you expect from the behaviors of your friends.Trust yourself that what you feel and see is accurate, then it becomes easier to trust and welcome new people who are formatted differently than either your parents or your friends who sound very similar to your parents' ways.
24
+ </s>
25
+ trainable params: 2,097,152 || all params: 264,507,392 || trainable%: 0.7929
26
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
27
+ warnings.warn(
28
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
29
+ warnings.warn(
30
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
31
+ warnings.warn(
32
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
33
+ warnings.warn(
34
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
35
+ warnings.warn(
36
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
37
+ warnings.warn(
38
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
39
+ warnings.warn(
40
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
41
+ warnings.warn(
42
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
43
+ warnings.warn(
44
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
45
+ warnings.warn(
46
+ [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
47
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
48
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
49
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
50
+ please provide the valuable advice for this comment.
51
+ i think i do not fit in this world
52
+ [/INST]
53
+ The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
54
+ Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
55
+ <s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
56
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
57
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
58
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
59
+ please provide the valuable advice for this comment.
60
+ i think i do not fit in this world
61
+ [/INST]
62
+ i think you are just trying to find your place in the world. It is okay to not fit in with everyone. It is okay to be different. It is okay to be yourself. It is okay to not be like everyone else. It is okay to be unique. It is okay to be you. It is okay to be happy. It is okay to be sad. It is okay to be angry. It is okay to be happy. It is okay to be sad. It is okay to be angry. It is okay to be happy.
63
+ [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
64
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
65
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
66
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
67
+ please provide the valuable advice for this comment.
68
+ I am not feeling good, my wife cheated on me , i think i should commit suicide
69
+ [/INST]
70
+ The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
71
+ Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
72
+ <s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
73
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
74
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
75
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
76
+ please provide the valuable advice for this comment.
77
+ I am not feeling good, my wife cheated on me , i think i should commit suicide
78
+ [/INST]
79
+ I'm sorry to hear that you are feeling this way. It is important to remember that you are not alone and that there are people who want to help you. I would encourage you to reach out to a therapist or counselor who can help you work through these feelings. If you are not comfortable doing that, there are also hotlines that you can call. The National Suicide Prevention Lifeline is 1-800-273-TALK (1-800-273-8255). You can also call the Crisis Text Line by texting 7
80
+ The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
81
+ Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
82
+ <s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
83
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
84
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
85
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
86
+ please provide the valuable advice for this comment.
87
+ I am not feeling good, my wife cheated on me , i think i should commit suicide
88
+ [/INST]
89
+ I'm sorry to hear that you are feeling this way. It is important to remember that you are not alone and that there are people who want to help you. I would encourage you to reach out to a therapist or counselor who can help you work through these feelings. If you are not comfortable doing that, there are also hotlines that you can call. The National Suicide Prevention Lifeline is 1-800-273-TALK (1-800-273-8255). You can also call the Crisis Text Line by texting 741741. I would also encourage you to consider joining a support group. Support groups can be a great resource for people who are dealing with the aftermath of an affair. You can search for a support group in your area by going to the website of
90
+ [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
91
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
92
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
93
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
94
+ please provide the valuable advice for this comment.
95
+ I am not feeling good, i think life is boring
96
+ [/INST]
97
+ The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
98
+ Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
99
+ <s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
100
+ state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
101
+ feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
102
+ Ensure your response is simple, relatable, encourages openness, and provides reassurance.
103
+ please provide the valuable advice for this comment.
104
+ I am not feeling good, i think life is boring
105
+ [/INST]
wandb/run-20240626_071628-smnm2aje/files/requirements.txt ADDED
@@ -0,0 +1,876 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.1.0
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.5
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.3
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.13.0
49
+ aiofiles==22.1.0
50
+ aiohttp==3.9.1
51
+ aioitertools==0.11.0
52
+ aiorwlock==1.3.0
53
+ aiosignal==1.3.1
54
+ aiosqlite==0.19.0
55
+ albumentations==1.4.0
56
+ alembic==1.13.1
57
+ altair==5.3.0
58
+ annotated-types==0.6.0
59
+ annotated-types==0.7.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.2.2
72
+ astropy-iers-data==0.2024.6.3.0.31.14
73
+ astropy==6.1.0
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ auto_gptq==0.7.1
81
+ autopep8==2.0.4
82
+ backoff==2.2.1
83
+ bayesian-optimization==1.4.3
84
+ beatrix_jupyterlab==2023.128.151533
85
+ beautifulsoup4==4.12.2
86
+ bitsandbytes==0.43.1
87
+ blake3==0.2.1
88
+ bleach==6.1.0
89
+ blessed==1.20.0
90
+ blinker==1.8.2
91
+ blis==0.7.10
92
+ blosc2==2.6.2
93
+ bokeh==3.4.1
94
+ boltons==23.1.1
95
+ boto3==1.26.100
96
+ botocore==1.34.106
97
+ bq_helper==0.4.1
98
+ bqplot==0.12.43
99
+ branca==0.7.2
100
+ brewer2mpl==1.4.1
101
+ brotlipy==0.7.0
102
+ cached-property==1.5.2
103
+ cachetools==4.2.4
104
+ cachetools==5.3.2
105
+ catalogue==2.0.10
106
+ catalyst==22.4
107
+ catboost==1.2.5
108
+ category-encoders==2.6.3
109
+ certifi==2024.2.2
110
+ cesium==0.12.1
111
+ cffi==1.16.0
112
+ charset-normalizer==3.3.2
113
+ chex==0.1.86
114
+ cleverhans==4.0.0
115
+ click-plugins==1.1.1
116
+ click==8.1.7
117
+ cligj==0.7.2
118
+ cloud-tpu-client==0.10
119
+ cloud-tpu-profiler==2.4.0
120
+ cloudpathlib==0.16.0
121
+ cloudpickle==2.2.1
122
+ cloudpickle==3.0.0
123
+ cmdstanpy==1.2.3
124
+ colorama==0.4.6
125
+ colorcet==3.1.0
126
+ coloredlogs==15.0.1
127
+ colorful==0.5.6
128
+ colorlog==6.8.2
129
+ colorlover==0.3.0
130
+ comm==0.2.1
131
+ conda-libmamba-solver==23.12.0
132
+ conda-package-handling==2.2.0
133
+ conda==24.5.0
134
+ conda_package_streaming==0.9.0
135
+ confection==0.1.4
136
+ contextily==1.6.0
137
+ contourpy==1.2.0
138
+ contourpy==1.2.1
139
+ convertdate==2.4.0
140
+ crcmod==1.7
141
+ cryptography==41.0.7
142
+ cuda-python==12.5.0
143
+ cudf==24.4.1
144
+ cufflinks==0.17.3
145
+ cuml==24.4.0
146
+ cupy==13.1.0
147
+ cycler==0.12.1
148
+ cymem==2.0.8
149
+ cytoolz==0.12.3
150
+ daal4py==2024.4.0
151
+ daal==2024.4.0
152
+ dacite==1.8.1
153
+ dask-cuda==24.4.0
154
+ dask-cudf==24.4.1
155
+ dask-expr==1.1.2
156
+ dask==2024.5.2
157
+ dataclasses-json==0.6.6
158
+ dataproc_jupyter_plugin==0.1.66
159
+ datasets==2.19.2
160
+ datashader==0.16.2
161
+ datatile==1.0.3
162
+ db-dtypes==1.2.0
163
+ deap==1.4.1
164
+ debugpy==1.8.0
165
+ decorator==5.1.1
166
+ deepdiff==7.0.1
167
+ defusedxml==0.7.1
168
+ deprecation==2.1.0
169
+ descartes==1.1.0
170
+ dill==0.3.8
171
+ dipy==1.9.0
172
+ distlib==0.3.8
173
+ distributed==2024.1.1
174
+ distro==1.9.0
175
+ dm-tree==0.1.8
176
+ docker-pycreds==0.4.0
177
+ docker==7.0.0
178
+ docopt==0.6.2
179
+ docstring-parser==0.15
180
+ docstring-to-markdown==0.15
181
+ docutils==0.21.2
182
+ earthengine-api==0.1.405
183
+ easydict==1.13
184
+ easyocr==1.7.1
185
+ ecos==2.0.13
186
+ eli5==0.13.0
187
+ emoji==2.12.1
188
+ en-core-web-lg==3.7.1
189
+ en-core-web-sm==3.7.1
190
+ entrypoints==0.4
191
+ ephem==4.1.5
192
+ esda==2.5.1
193
+ essentia==2.1b6.dev1110
194
+ et-xmlfile==1.1.0
195
+ etils==1.6.0
196
+ exceptiongroup==1.2.0
197
+ executing==2.0.1
198
+ explainable-ai-sdk==1.3.3
199
+ fastai==2.7.15
200
+ fastapi==0.108.0
201
+ fastavro==1.9.3
202
+ fastcore==1.5.43
203
+ fastdownload==0.0.7
204
+ fasteners==0.19
205
+ fastjsonschema==2.19.1
206
+ fastprogress==1.0.3
207
+ fastrlock==0.8.2
208
+ fasttext==0.9.2
209
+ feather-format==0.4.1
210
+ featuretools==1.31.0
211
+ filelock==3.13.1
212
+ fiona==1.9.6
213
+ fitter==1.7.0
214
+ flake8==7.0.0
215
+ flashtext==2.7
216
+ flatbuffers==23.5.26
217
+ flax==0.8.4
218
+ folium==0.16.0
219
+ fonttools==4.47.0
220
+ fonttools==4.53.0
221
+ fqdn==1.5.1
222
+ frozendict==2.4.4
223
+ frozenlist==1.4.1
224
+ fsspec==2024.3.1
225
+ fsspec==2024.5.0
226
+ funcy==2.0
227
+ fury==0.10.0
228
+ future==1.0.0
229
+ fuzzywuzzy==0.18.0
230
+ gast==0.5.4
231
+ gatspy==0.3
232
+ gcsfs==2024.3.1
233
+ gekko==1.1.3
234
+ gensim==4.3.2
235
+ geographiclib==2.0
236
+ geojson==3.1.0
237
+ geopandas==0.14.4
238
+ geoplot==0.5.1
239
+ geopy==2.4.1
240
+ geoviews==1.12.0
241
+ ggplot==0.11.5
242
+ giddy==2.3.5
243
+ gitdb==4.0.11
244
+ google-ai-generativelanguage==0.6.4
245
+ google-api-core==2.11.1
246
+ google-api-core==2.19.0
247
+ google-api-python-client==2.131.0
248
+ google-apitools==0.5.31
249
+ google-auth-httplib2==0.2.0
250
+ google-auth-oauthlib==1.2.0
251
+ google-auth==2.26.1
252
+ google-cloud-aiplatform==0.6.0a1
253
+ google-cloud-artifact-registry==1.10.0
254
+ google-cloud-automl==1.0.1
255
+ google-cloud-bigquery==2.34.4
256
+ google-cloud-bigtable==1.7.3
257
+ google-cloud-core==2.4.1
258
+ google-cloud-datastore==2.19.0
259
+ google-cloud-dlp==3.14.0
260
+ google-cloud-jupyter-config==0.0.5
261
+ google-cloud-language==2.13.3
262
+ google-cloud-monitoring==2.18.0
263
+ google-cloud-pubsub==2.19.0
264
+ google-cloud-pubsublite==1.9.0
265
+ google-cloud-recommendations-ai==0.7.1
266
+ google-cloud-resource-manager==1.11.0
267
+ google-cloud-spanner==3.40.1
268
+ google-cloud-storage==1.44.0
269
+ google-cloud-translate==3.12.1
270
+ google-cloud-videointelligence==2.13.3
271
+ google-cloud-vision==2.8.0
272
+ google-crc32c==1.5.0
273
+ google-generativeai==0.6.0
274
+ google-pasta==0.2.0
275
+ google-resumable-media==2.7.0
276
+ googleapis-common-protos==1.62.0
277
+ gplearn==0.4.2
278
+ gpustat==1.0.0
279
+ gpxpy==1.6.2
280
+ graphviz==0.20.3
281
+ greenlet==3.0.3
282
+ grpc-google-iam-v1==0.12.7
283
+ grpcio-status==1.48.1
284
+ grpcio-status==1.48.2
285
+ grpcio==1.59.3
286
+ grpcio==1.60.0
287
+ gviz-api==1.10.0
288
+ gym-notices==0.0.8
289
+ gym==0.26.2
290
+ gymnasium==0.29.0
291
+ h11==0.14.0
292
+ h2o==3.46.0.2
293
+ h5netcdf==1.3.0
294
+ h5py==3.10.0
295
+ haversine==2.8.1
296
+ hdfs==2.7.3
297
+ hep-ml==0.7.2
298
+ hijri-converter==2.3.1
299
+ hmmlearn==0.3.2
300
+ holidays==0.24
301
+ holoviews==1.18.3
302
+ hpsklearn==0.1.0
303
+ html5lib==1.1
304
+ htmlmin==0.1.12
305
+ httpcore==1.0.5
306
+ httplib2==0.21.0
307
+ httptools==0.6.1
308
+ httpx==0.27.0
309
+ huggingface-hub==0.23.2
310
+ humanfriendly==10.0
311
+ hunspell==0.5.5
312
+ hydra-slayer==0.5.0
313
+ hyperopt==0.2.7
314
+ hypertools==0.8.0
315
+ idna==3.6
316
+ igraph==0.11.5
317
+ imagecodecs==2024.6.1
318
+ imageio==2.33.1
319
+ imbalanced-learn==0.12.3
320
+ imgaug==0.4.0
321
+ importlib-metadata==6.11.0
322
+ importlib-metadata==7.0.1
323
+ importlib-resources==6.1.1
324
+ inequality==1.0.1
325
+ iniconfig==2.0.0
326
+ ipydatawidgets==4.3.5
327
+ ipykernel==6.28.0
328
+ ipyleaflet==0.19.1
329
+ ipympl==0.7.0
330
+ ipython-genutils==0.2.0
331
+ ipython-genutils==0.2.0
332
+ ipython-sql==0.5.0
333
+ ipython==8.20.0
334
+ ipyvolume==0.6.3
335
+ ipyvue==1.11.1
336
+ ipyvuetify==1.9.4
337
+ ipywebrtc==0.6.0
338
+ ipywidgets==7.7.1
339
+ isoduration==20.11.0
340
+ isort==5.13.2
341
+ isoweek==1.3.3
342
+ itsdangerous==2.2.0
343
+ jaraco.classes==3.3.0
344
+ jax-jumpy==1.0.0
345
+ jax==0.4.26
346
+ jaxlib==0.4.26.dev20240504
347
+ jedi==0.19.1
348
+ jeepney==0.8.0
349
+ jieba==0.42.1
350
+ jmespath==1.0.1
351
+ joblib==1.4.2
352
+ json5==0.9.14
353
+ jsonpatch==1.33
354
+ jsonpointer==2.4
355
+ jsonschema-specifications==2023.12.1
356
+ jsonschema==4.20.0
357
+ jupyter-console==6.6.3
358
+ jupyter-events==0.9.0
359
+ jupyter-http-over-ws==0.0.8
360
+ jupyter-leaflet==0.19.1
361
+ jupyter-lsp==1.5.1
362
+ jupyter-server-mathjax==0.2.6
363
+ jupyter-ydoc==0.2.5
364
+ jupyter_client==7.4.9
365
+ jupyter_client==8.6.0
366
+ jupyter_core==5.7.1
367
+ jupyter_server==2.12.5
368
+ jupyter_server_fileid==0.9.1
369
+ jupyter_server_proxy==4.1.0
370
+ jupyter_server_terminals==0.5.1
371
+ jupyter_server_ydoc==0.8.0
372
+ jupyterlab-lsp==5.1.0
373
+ jupyterlab-widgets==3.0.9
374
+ jupyterlab==4.2.1
375
+ jupyterlab_git==0.44.0
376
+ jupyterlab_pygments==0.3.0
377
+ jupyterlab_server==2.27.2
378
+ jupytext==1.16.0
379
+ kaggle-environments==1.14.11
380
+ kaggle==1.6.14
381
+ kagglehub==0.2.5
382
+ keras-cv==0.9.0
383
+ keras-nlp==0.12.1
384
+ keras-tuner==1.4.6
385
+ keras==3.3.3
386
+ kernels-mixer==0.0.7
387
+ keyring==24.3.0
388
+ keyrings.google-artifactregistry-auth==1.1.2
389
+ kfp-pipeline-spec==0.2.2
390
+ kfp-server-api==2.0.5
391
+ kfp==2.5.0
392
+ kiwisolver==1.4.5
393
+ kmapper==2.0.1
394
+ kmodes==0.12.2
395
+ korean-lunar-calendar==0.3.1
396
+ kornia==0.7.2
397
+ kornia_rs==0.1.3
398
+ kt-legacy==1.0.5
399
+ kubernetes==26.1.0
400
+ langcodes==3.4.0
401
+ langid==1.1.6
402
+ language_data==1.2.0
403
+ lazy_loader==0.3
404
+ learntools==0.3.4
405
+ leven==1.0.4
406
+ libclang==16.0.6
407
+ libmambapy==1.5.8
408
+ libpysal==4.9.2
409
+ librosa==0.10.2.post1
410
+ lightgbm==4.2.0
411
+ lightning-utilities==0.11.2
412
+ lime==0.2.0.1
413
+ line_profiler==4.1.3
414
+ linkify-it-py==2.0.3
415
+ llvmlite==0.41.1
416
+ llvmlite==0.42.0
417
+ lml==0.1.0
418
+ locket==1.0.0
419
+ loguru==0.7.2
420
+ lxml==5.2.2
421
+ lz4==4.3.3
422
+ mamba==1.5.8
423
+ mapclassify==2.6.1
424
+ marisa-trie==1.1.0
425
+ markdown-it-py==3.0.0
426
+ marshmallow==3.21.2
427
+ matplotlib-inline==0.1.6
428
+ matplotlib-venn==0.11.10
429
+ matplotlib==3.7.5
430
+ matplotlib==3.8.4
431
+ mccabe==0.7.0
432
+ mdit-py-plugins==0.4.0
433
+ mdurl==0.1.2
434
+ memory-profiler==0.61.0
435
+ menuinst==2.0.1
436
+ mercantile==1.2.1
437
+ mgwr==2.2.1
438
+ missingno==0.5.2
439
+ mistune==0.8.4
440
+ mizani==0.11.4
441
+ ml-dtypes==0.2.0
442
+ mlcrate==0.2.0
443
+ mlens==0.2.3
444
+ mlxtend==0.23.1
445
+ mne==1.7.0
446
+ mnist==0.2.2
447
+ momepy==0.7.0
448
+ more-itertools==10.2.0
449
+ mpld3==0.5.10
450
+ mpmath==1.3.0
451
+ msgpack==1.0.7
452
+ msgpack==1.0.8
453
+ multidict==6.0.4
454
+ multimethod==1.10
455
+ multipledispatch==1.0.0
456
+ multiprocess==0.70.16
457
+ munkres==1.1.4
458
+ murmurhash==1.0.10
459
+ mypy-extensions==1.0.0
460
+ namex==0.0.8
461
+ nb-conda-kernels==2.3.1
462
+ nb_conda==2.2.1
463
+ nbclassic==1.0.0
464
+ nbclient==0.5.13
465
+ nbclient==0.9.0
466
+ nbconvert==6.4.5
467
+ nbdime==3.2.0
468
+ nbformat==5.9.2
469
+ ndindex==1.8
470
+ nest-asyncio==1.5.8
471
+ networkx==3.2.1
472
+ nibabel==5.2.1
473
+ nilearn==0.10.4
474
+ ninja==1.11.1.1
475
+ nltk==3.2.4
476
+ nose==1.3.7
477
+ notebook==6.5.4
478
+ notebook==6.5.6
479
+ notebook_executor==0.2
480
+ notebook_shim==0.2.3
481
+ numba==0.58.1
482
+ numba==0.59.1
483
+ numexpr==2.10.0
484
+ numpy==1.26.4
485
+ nvidia-ml-py==11.495.46
486
+ nvtx==0.2.10
487
+ oauth2client==4.1.3
488
+ oauthlib==3.2.2
489
+ objsize==0.6.1
490
+ odfpy==1.4.1
491
+ olefile==0.47
492
+ onnx==1.16.1
493
+ opencensus-context==0.1.3
494
+ opencensus==0.11.4
495
+ opencv-contrib-python==4.10.0.82
496
+ opencv-python-headless==4.10.0.82
497
+ opencv-python==4.10.0.82
498
+ openpyxl==3.1.3
499
+ openslide-python==1.3.1
500
+ opentelemetry-api==1.22.0
501
+ opentelemetry-exporter-otlp-proto-common==1.22.0
502
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
503
+ opentelemetry-exporter-otlp-proto-http==1.22.0
504
+ opentelemetry-exporter-otlp==1.22.0
505
+ opentelemetry-proto==1.22.0
506
+ opentelemetry-sdk==1.22.0
507
+ opentelemetry-semantic-conventions==0.43b0
508
+ opt-einsum==3.3.0
509
+ optax==0.2.2
510
+ optimum==1.20.0
511
+ optree==0.11.0
512
+ optuna==3.6.1
513
+ orbax-checkpoint==0.5.15
514
+ ordered-set==4.1.0
515
+ orjson==3.9.10
516
+ ortools==9.4.1874
517
+ osmnx==1.9.3
518
+ overrides==7.4.0
519
+ packaging==21.3
520
+ pandas-datareader==0.10.0
521
+ pandas-profiling==3.6.6
522
+ pandas-summary==0.2.0
523
+ pandas==2.2.1
524
+ pandas==2.2.2
525
+ pandasql==0.7.3
526
+ pandocfilters==1.5.0
527
+ panel==1.4.4
528
+ papermill==2.5.0
529
+ param==2.1.0
530
+ parso==0.8.3
531
+ partd==1.4.2
532
+ path.py==12.5.0
533
+ path==16.14.0
534
+ pathos==0.3.2
535
+ pathy==0.10.3
536
+ patsy==0.5.6
537
+ pdf2image==1.17.0
538
+ peft==0.11.1
539
+ pettingzoo==1.24.0
540
+ pexpect==4.8.0
541
+ pexpect==4.9.0
542
+ phik==0.12.4
543
+ pickleshare==0.7.5
544
+ pillow==10.3.0
545
+ pip==23.3.2
546
+ pkgutil_resolve_name==1.3.10
547
+ platformdirs==3.11.0
548
+ platformdirs==4.1.0
549
+ plotly-express==0.4.1
550
+ plotly==5.18.0
551
+ plotnine==0.13.6
552
+ pluggy==1.5.0
553
+ pointpats==2.4.0
554
+ polars==0.20.31
555
+ polyglot==16.7.4
556
+ pooch==1.8.1
557
+ pox==0.3.4
558
+ ppca==0.0.4
559
+ ppft==1.7.6.8
560
+ preprocessing==0.1.13
561
+ preshed==3.0.9
562
+ prettytable==3.9.0
563
+ progressbar2==4.4.2
564
+ prometheus-client==0.19.0
565
+ promise==2.3
566
+ prompt-toolkit==3.0.42
567
+ prompt-toolkit==3.0.43
568
+ prophet==1.1.1
569
+ proto-plus==1.23.0
570
+ protobuf==3.20.3
571
+ protobuf==4.24.4
572
+ psutil==5.9.3
573
+ psutil==5.9.7
574
+ ptyprocess==0.7.0
575
+ pudb==2024.1
576
+ pure-eval==0.2.2
577
+ py-cpuinfo==9.0.0
578
+ py-spy==0.3.14
579
+ py4j==0.10.9.7
580
+ pyLDAvis==3.4.1
581
+ pyOpenSSL==23.3.0
582
+ pyaml==24.4.0
583
+ pyarrow-hotfix==0.6
584
+ pyarrow==14.0.2
585
+ pyasn1-modules==0.3.0
586
+ pyasn1==0.5.1
587
+ pybind11==2.12.0
588
+ pyclipper==1.3.0.post5
589
+ pycodestyle==2.11.1
590
+ pycosat==0.6.6
591
+ pycparser==2.21
592
+ pycryptodome==3.20.0
593
+ pyct==0.5.0
594
+ pycuda==2024.1
595
+ pydantic==2.5.3
596
+ pydantic==2.7.2
597
+ pydantic_core==2.14.6
598
+ pydantic_core==2.18.3
599
+ pydegensac==0.1.2
600
+ pydicom==2.4.4
601
+ pydocstyle==6.3.0
602
+ pydot==1.4.2
603
+ pydub==0.25.1
604
+ pyemd==1.0.0
605
+ pyerfa==2.0.1.4
606
+ pyexcel-io==0.6.6
607
+ pyexcel-ods==0.6.0
608
+ pyflakes==3.2.0
609
+ pygltflib==1.16.2
610
+ pykalman==0.9.7
611
+ pylibraft==24.4.0
612
+ pylint==3.2.2
613
+ pymc3==3.11.4
614
+ pymongo==3.13.0
615
+ pynndescent==0.5.12
616
+ pynvjitlink==0.2.3
617
+ pynvml==11.4.1
618
+ pynvrtc==9.2
619
+ pyparsing==3.1.1
620
+ pyparsing==3.1.2
621
+ pypdf==4.2.0
622
+ pyproj==3.6.1
623
+ pysal==24.1
624
+ pyshp==2.3.1
625
+ pytesseract==0.3.10
626
+ pytest==8.2.1
627
+ python-bidi==0.4.2
628
+ python-dateutil==2.9.0.post0
629
+ python-dotenv==1.0.0
630
+ python-json-logger==2.0.7
631
+ python-louvain==0.16
632
+ python-lsp-jsonrpc==1.1.2
633
+ python-lsp-server==1.11.0
634
+ python-slugify==8.0.4
635
+ python-utils==3.8.2
636
+ pythreejs==2.4.2
637
+ pytoolconfig==1.3.1
638
+ pytools==2024.1.3
639
+ pytorch-ignite==0.5.0.post2
640
+ pytorch-lightning==2.2.5
641
+ pytz==2023.3.post1
642
+ pytz==2024.1
643
+ pyu2f==0.1.5
644
+ pyviz_comms==3.0.2
645
+ pyzmq==24.0.1
646
+ pyzmq==25.1.2
647
+ qgrid==1.3.1
648
+ qtconsole==5.5.2
649
+ quantecon==0.7.2
650
+ qudida==0.0.4
651
+ raft-dask==24.4.0
652
+ rapids-dask-dependency==24.4.1a0
653
+ rasterio==1.3.10
654
+ rasterstats==0.19.0
655
+ ray-cpp==2.9.0
656
+ ray==2.9.0
657
+ referencing==0.32.1
658
+ regex==2023.12.25
659
+ requests-oauthlib==1.3.1
660
+ requests-toolbelt==0.10.1
661
+ requests==2.32.3
662
+ retrying==1.3.3
663
+ retrying==1.3.4
664
+ rfc3339-validator==0.1.4
665
+ rfc3986-validator==0.1.1
666
+ rgf-python==3.12.0
667
+ rich-click==1.8.2
668
+ rich==13.7.0
669
+ rich==13.7.1
670
+ rmm==24.4.0
671
+ rope==1.13.0
672
+ rouge==1.0.1
673
+ rpds-py==0.16.2
674
+ rsa==4.9
675
+ ruamel-yaml-conda==0.15.100
676
+ ruamel.yaml.clib==0.2.7
677
+ ruamel.yaml==0.18.5
678
+ s2sphere==0.2.5
679
+ s3fs==2024.3.1
680
+ s3transfer==0.6.2
681
+ safetensors==0.4.3
682
+ scattertext==0.1.19
683
+ scikit-image==0.22.0
684
+ scikit-learn-intelex==2024.4.0
685
+ scikit-learn==1.2.2
686
+ scikit-multilearn==0.2.0
687
+ scikit-optimize==0.10.1
688
+ scikit-plot==0.3.7
689
+ scikit-surprise==1.1.4
690
+ scipy==1.11.4
691
+ scipy==1.13.1
692
+ seaborn==0.12.2
693
+ segment_anything==1.0
694
+ segregation==2.5
695
+ semver==3.0.2
696
+ sentencepiece==0.2.0
697
+ sentry-sdk==2.3.1
698
+ setproctitle==1.3.3
699
+ setuptools-git==1.2
700
+ setuptools-scm==8.1.0
701
+ setuptools==69.0.3
702
+ shap==0.44.1
703
+ shapely==2.0.4
704
+ shellingham==1.5.4
705
+ simpervisor==1.0.0
706
+ simplejson==3.19.2
707
+ six==1.16.0
708
+ sklearn-pandas==2.2.0
709
+ slicer==0.0.7
710
+ smart-open==6.4.0
711
+ smmap==5.0.1
712
+ sniffio==1.3.0
713
+ snowballstemmer==2.2.0
714
+ snuggs==1.4.7
715
+ sortedcontainers==2.4.0
716
+ soundfile==0.12.1
717
+ soupsieve==2.5
718
+ soxr==0.3.7
719
+ spacy-legacy==3.0.12
720
+ spacy-loggers==1.0.5
721
+ spacy==3.7.3
722
+ spaghetti==1.7.5.post1
723
+ spectral==0.23.1
724
+ spglm==1.1.0
725
+ sphinx-rtd-theme==0.2.4
726
+ spint==1.0.7
727
+ splot==1.1.5.post1
728
+ spopt==0.6.0
729
+ spreg==1.4.2
730
+ spvcm==0.3.0
731
+ sqlparse==0.4.4
732
+ squarify==0.4.3
733
+ srsly==2.4.8
734
+ stable-baselines3==2.1.0
735
+ stack-data==0.6.2
736
+ stack-data==0.6.3
737
+ stanio==0.5.0
738
+ starlette==0.32.0.post1
739
+ statsmodels==0.14.1
740
+ stemming==1.0.1
741
+ stop-words==2018.7.23
742
+ stopit==1.1.2
743
+ stumpy==1.12.0
744
+ sympy==1.12.1
745
+ tables==3.9.2
746
+ tabulate==0.9.0
747
+ tangled-up-in-unicode==0.2.0
748
+ tbb==2021.12.0
749
+ tblib==3.0.0
750
+ tenacity==8.2.3
751
+ tensorboard-data-server==0.7.2
752
+ tensorboard-plugin-profile==2.15.0
753
+ tensorboard==2.15.1
754
+ tensorboardX==2.6.2.2
755
+ tensorflow-cloud==0.1.16
756
+ tensorflow-datasets==4.9.4
757
+ tensorflow-decision-forests==1.8.1
758
+ tensorflow-estimator==2.15.0
759
+ tensorflow-hub==0.16.1
760
+ tensorflow-io-gcs-filesystem==0.35.0
761
+ tensorflow-io==0.35.0
762
+ tensorflow-metadata==0.14.0
763
+ tensorflow-probability==0.23.0
764
+ tensorflow-serving-api==2.14.1
765
+ tensorflow-text==2.15.0
766
+ tensorflow-transform==0.14.0
767
+ tensorflow==2.15.0
768
+ tensorstore==0.1.60
769
+ termcolor==2.4.0
770
+ terminado==0.18.0
771
+ testpath==0.6.0
772
+ text-unidecode==1.3
773
+ textblob==0.18.0.post0
774
+ texttable==1.7.0
775
+ tf_keras==2.15.1
776
+ tfp-nightly==0.24.0.dev0
777
+ thinc==8.2.3
778
+ threadpoolctl==3.2.0
779
+ tifffile==2023.12.9
780
+ timm==1.0.3
781
+ tinycss2==1.2.1
782
+ tobler==0.11.2
783
+ tokenizers==0.19.1
784
+ toml==0.10.2
785
+ tomli==2.0.1
786
+ tomlkit==0.12.5
787
+ toolz==0.12.1
788
+ torch==2.1.2
789
+ torchaudio==2.1.2
790
+ torchdata==0.7.1
791
+ torchinfo==1.8.0
792
+ torchmetrics==1.4.0.post0
793
+ torchtext==0.16.2
794
+ torchvision==0.16.2
795
+ tornado==6.3.3
796
+ tqdm==4.66.4
797
+ traceml==1.0.8
798
+ traitlets==5.9.0
799
+ traittypes==0.2.1
800
+ transformers==4.41.2
801
+ treelite==4.1.2
802
+ truststore==0.8.0
803
+ trx-python==0.2.9
804
+ tsfresh==0.20.2
805
+ typeguard==4.1.5
806
+ typer==0.9.0
807
+ typer==0.9.4
808
+ types-python-dateutil==2.8.19.20240106
809
+ typing-inspect==0.9.0
810
+ typing-utils==0.1.0
811
+ typing_extensions==4.9.0
812
+ tzdata==2023.4
813
+ tzdata==2024.1
814
+ uc-micro-py==1.0.3
815
+ ucx-py==0.37.0
816
+ ujson==5.10.0
817
+ umap-learn==0.5.6
818
+ unicodedata2==15.1.0
819
+ update-checker==0.18.0
820
+ uri-template==1.3.0
821
+ uritemplate==3.0.1
822
+ urllib3==1.26.18
823
+ urllib3==2.1.0
824
+ urwid==2.6.12
825
+ urwid_readline==0.14
826
+ uvicorn==0.25.0
827
+ uvloop==0.19.0
828
+ vaex-astro==0.9.3
829
+ vaex-core==4.17.1
830
+ vaex-hdf5==0.14.1
831
+ vaex-jupyter==0.8.2
832
+ vaex-ml==0.18.3
833
+ vaex-server==0.9.0
834
+ vaex-viz==0.5.4
835
+ vaex==4.17.0
836
+ vec_noise==1.1.4
837
+ vecstack==0.4.0
838
+ virtualenv==20.21.0
839
+ visions==0.7.5
840
+ vowpalwabbit==9.9.0
841
+ vtk==9.3.0
842
+ wandb==0.17.0
843
+ wasabi==1.1.2
844
+ watchfiles==0.21.0
845
+ wavio==0.0.9
846
+ wcwidth==0.2.13
847
+ weasel==0.3.4
848
+ webcolors==1.13
849
+ webencodings==0.5.1
850
+ websocket-client==1.7.0
851
+ websockets==12.0
852
+ wfdb==4.1.2
853
+ whatthepatch==1.0.5
854
+ wheel==0.42.0
855
+ widgetsnbextension==3.6.6
856
+ witwidget==1.8.1
857
+ woodwork==0.31.0
858
+ wordcloud==1.9.3
859
+ wordsegment==1.3.1
860
+ wrapt==1.14.1
861
+ xarray-einstats==0.7.0
862
+ xarray==2024.5.0
863
+ xgboost==2.0.3
864
+ xvfbwrapper==0.2.9
865
+ xxhash==3.4.1
866
+ xyzservices==2024.4.0
867
+ y-py==0.6.2
868
+ yapf==0.40.2
869
+ yarl==1.9.3
870
+ yarl==1.9.4
871
+ ydata-profiling==4.6.4
872
+ yellowbrick==1.5
873
+ ypy-websocket==0.8.4
874
+ zict==3.0.0
875
+ zipp==3.17.0
876
+ zstandard==0.19.0
wandb/run-20240626_071628-smnm2aje/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-06-26T07:16:29.539044",
5
+ "startedAt": "2024-06-26T07:16:28.965434",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "adc95cf38b20",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.146,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.146,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.146,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.146,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.146,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5693.895751953125
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357563018798828
65
+ }
66
+ }
wandb/run-20240626_071628-smnm2aje/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.18, "train/grad_norm": 0.583970308303833, "train/learning_rate": 0.0, "train/epoch": 9.411764705882353, "train/global_step": 120, "_timestamp": 1719391099.5724134, "_runtime": 4910.594539403915, "_step": 21, "eval/loss": 1.2984201908111572, "eval/runtime": 127.8842, "eval/samples_per_second": 1.572, "eval/steps_per_second": 0.399, "train_runtime": 4600.4395, "train_samples_per_second": 0.437, "train_steps_per_second": 0.026, "total_flos": 775966301356032.0, "train_loss": 1.6367768367131552}
wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240626_071628-smnm2aje/logs/debug.log ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
10
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
11
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
12
+ 2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
13
+ 2024-06-26 07:16:28,967 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
14
+ Traceback (most recent call last):
15
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
16
+ exec(code_obj, self.user_global_ns, self.user_ns)
17
+ File "/tmp/ipykernel_34/2014566126.py", line 10, in <module>
18
+ trainer.train()
19
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
20
+ return inner_training_loop(
21
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
22
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
23
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
24
+ return self.call_event("on_train_begin", args, state, control)
25
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
26
+ result = getattr(callback, event)(
27
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
28
+ self.setup(args, state, model, **kwargs)
29
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
30
+ self._wandb.init(
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
32
+ wandb._sentry.reraise(e)
33
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
34
+ raise exc.with_traceback(sys.exc_info()[2])
35
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
36
+ wi.setup(kwargs)
37
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
38
+ wandb_login._login(
39
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
40
+ wlogin.prompt_api_key()
41
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
42
+ key, status = self._prompt_api_key()
43
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
44
+ key = apikey.prompt_api_key(
45
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
46
+ key = input_callback(api_ask).strip()
47
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
48
+ value = prompt_func(prompt)
49
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
50
+ raise Abort() from None
51
+ click.exceptions.Abort
52
+ 2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
53
+ Traceback (most recent call last):
54
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
55
+ exec(code_obj, self.user_global_ns, self.user_ns)
56
+ File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
57
+ trainer.train()
58
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
59
+ return inner_training_loop(
60
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
61
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
62
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
63
+ return self.call_event("on_train_begin", args, state, control)
64
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
65
+ result = getattr(callback, event)(
66
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
67
+ self.setup(args, state, model, **kwargs)
68
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
69
+ self._wandb.init(
70
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
71
+ wandb._sentry.reraise(e)
72
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
73
+ raise exc.with_traceback(sys.exc_info()[2])
74
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
75
+ wi.setup(kwargs)
76
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
77
+ wandb_login._login(
78
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
79
+ wlogin.prompt_api_key()
80
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
81
+ key, status = self._prompt_api_key()
82
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
83
+ key = apikey.prompt_api_key(
84
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
85
+ key = input_callback(api_ask).strip()
86
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
87
+ value = prompt_func(prompt)
88
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
89
+ raise Abort() from None
90
+ click.exceptions.Abort
91
+ 2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
92
+ Traceback (most recent call last):
93
+ File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
94
+ exec(code_obj, self.user_global_ns, self.user_ns)
95
+ File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
96
+ trainer.train()
97
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
98
+ return inner_training_loop(
99
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
100
+ self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
101
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
102
+ return self.call_event("on_train_begin", args, state, control)
103
+ File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
104
+ result = getattr(callback, event)(
105
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
106
+ self.setup(args, state, model, **kwargs)
107
+ File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
108
+ self._wandb.init(
109
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
110
+ wandb._sentry.reraise(e)
111
+ File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
112
+ raise exc.with_traceback(sys.exc_info()[2])
113
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
114
+ wi.setup(kwargs)
115
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
116
+ wandb_login._login(
117
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
118
+ wlogin.prompt_api_key()
119
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
120
+ key, status = self._prompt_api_key()
121
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
122
+ key = apikey.prompt_api_key(
123
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
124
+ key = input_callback(api_ask).strip()
125
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
126
+ value = prompt_func(prompt)
127
+ File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
128
+ raise Abort() from None
129
+ click.exceptions.Abort
130
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug.log
131
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log
132
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78720aaa5b40>
133
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
134
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
135
+ config: {}
136
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():610] starting backend
137
+ 2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
138
+ 2024-06-26 07:16:28,971 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
139
+ 2024-06-26 07:16:28,977 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
140
+ 2024-06-26 07:16:28,989 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
141
+ 2024-06-26 07:16:29,303 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
142
+ 2024-06-26 07:16:29,306 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
143
+ 2024-06-26 07:16:29,429 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
144
+ 2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
145
+
146
+ 2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
147
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
148
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
149
+ 2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
150
+ 2024-06-26 07:16:45,663 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
151
+ 2024-06-26 07:16:45,672 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
152
+ 2024-06-26 07:16:45,678 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-16-15_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
153
+ 2024-06-26 07:18:26,851 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
154
+ 2024-06-26 07:18:26,851 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
155
+ 2024-06-26 07:18:50,892 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
156
+ 2024-06-26 07:18:50,915 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
157
+ 2024-06-26 07:18:50,915 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
158
+ 2024-06-26 07:18:51,909 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
159
+ 2024-06-26 07:18:51,913 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
160
+ 2024-06-26 07:18:51,913 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
161
+ 2024-06-26 07:18:52,853 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
162
+ 2024-06-26 07:18:52,855 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
163
+ 2024-06-26 07:18:52,855 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
164
+ 2024-06-26 07:18:53,667 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
165
+ 2024-06-26 07:18:53,670 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
166
+ 2024-06-26 07:18:53,670 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
167
+ 2024-06-26 07:18:54,691 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
168
+ 2024-06-26 07:18:54,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
169
+ 2024-06-26 07:18:54,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
170
+ 2024-06-26 07:18:55,617 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
171
+ 2024-06-26 07:18:55,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
172
+ 2024-06-26 07:18:55,682 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
173
+ 2024-06-26 07:18:56,512 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
174
+ 2024-06-26 07:18:56,513 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
175
+ 2024-06-26 07:18:56,513 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
176
+ 2024-06-26 07:18:57,999 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
177
+ 2024-06-26 07:18:58,001 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
178
+ 2024-06-26 07:18:58,001 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
179
+ 2024-06-26 07:18:58,997 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
180
+ 2024-06-26 07:18:59,009 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
181
+ 2024-06-26 07:18:59,009 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
182
+ 2024-06-26 07:19:00,090 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
183
+ 2024-06-26 07:19:00,092 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
184
+ 2024-06-26 07:19:00,092 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
185
+ 2024-06-26 07:19:01,722 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
186
+ 2024-06-26 07:19:01,723 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
187
+ 2024-06-26 07:19:01,723 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
188
+ 2024-06-26 07:19:03,098 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
189
+ 2024-06-26 07:19:03,117 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
190
+ 2024-06-26 07:19:03,117 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
191
+ 2024-06-26 07:19:04,117 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
192
+ 2024-06-26 07:19:04,212 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
193
+ 2024-06-26 07:19:04,212 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
194
+ 2024-06-26 07:19:05,065 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
195
+ 2024-06-26 07:19:05,097 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
196
+ 2024-06-26 07:19:05,097 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
197
+ 2024-06-26 07:19:06,209 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
198
+ 2024-06-26 07:19:06,220 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
199
+ 2024-06-26 07:19:06,220 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
200
+ 2024-06-26 07:19:07,300 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
201
+ 2024-06-26 07:19:07,961 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-19-05_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
202
+ 2024-06-26 08:35:48,386 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
203
+ 2024-06-26 08:35:48,387 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
204
+ 2024-06-26 08:36:11,677 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
205
+ 2024-06-26 08:38:19,575 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
206
+ 2024-06-26 08:38:19,575 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
207
+ 2024-06-26 08:39:55,100 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
208
+ 2024-06-26 08:39:55,103 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
209
+ 2024-06-26 08:39:55,103 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
210
+ 2024-06-26 08:40:32,017 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
211
+ 2024-06-26 08:43:02,593 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
212
+ 2024-06-26 08:43:02,594 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
213
+ 2024-06-26 08:45:10,959 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
214
+ 2024-06-26 08:45:10,963 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
215
+ 2024-06-26 08:45:10,963 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
216
+ 2024-06-26 08:45:13,384 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
217
+ 2024-06-26 08:47:43,994 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
218
+ 2024-06-26 08:47:43,994 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
219
+ 2024-06-26 08:49:58,760 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
220
+ 2024-06-26 08:53:33,771 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
221
+ 2024-06-26 08:53:33,771 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
222
+ 2024-06-26 08:54:32,055 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
223
+ 2024-06-26 08:54:32,059 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
224
+ 2024-06-26 08:54:32,059 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
225
+ 2024-06-26 08:54:32,754 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
226
+ 2024-06-26 08:58:07,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
227
+ 2024-06-26 08:58:07,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
228
+ 2024-06-26 09:12:46,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
229
+ 2024-06-26 09:12:46,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
230
+ 2024-06-26 09:12:46,669 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
231
+ 2024-06-26 09:16:22,037 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
232
+ 2024-06-26 09:16:22,038 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
233
+ 2024-06-26 09:16:22,038 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
234
+ 2024-06-26 09:16:35,228 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
235
+ 2024-06-26 09:16:36,438 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
236
+ 2024-06-26 09:16:36,438 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
237
+ 2024-06-26 09:18:25,016 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
238
+ 2024-06-26 09:18:26,213 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
239
+ 2024-06-26 09:18:26,214 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
240
+ 2024-06-26 09:21:21,819 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
241
+ 2024-06-26 09:21:21,842 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
242
+ 2024-06-26 09:21:21,842 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
243
+ 2024-06-26 09:21:36,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
244
+ 2024-06-26 09:21:36,633 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
245
+ 2024-06-26 09:21:36,634 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
246
+ 2024-06-26 09:21:37,142 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240626_071628-smnm2aje/run-smnm2aje.wandb ADDED
Binary file (212 kB). View file