mohityadav
commited on
Commit
•
a184792
1
Parent(s):
d393d46
mohityadav/mental-health-advisor-gpt
Browse files- README.md +72 -0
- adapter_config.json +28 -0
- adapter_model.safetensors +3 -0
- runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385345.adc95cf38b20.34.0 +3 -0
- runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385423.adc95cf38b20.34.1 +3 -0
- runs/Jun26_07-07-26_adc95cf38b20/events.out.tfevents.1719385648.adc95cf38b20.34.2 +3 -0
- runs/Jun26_07-16-15_adc95cf38b20/events.out.tfevents.1719386177.adc95cf38b20.34.3 +3 -0
- runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719386347.adc95cf38b20.34.4 +3 -0
- runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719391099.adc95cf38b20.34.5 +3 -0
- training_args.bin +3 -0
- wandb/debug-internal.log +0 -0
- wandb/debug.log +246 -0
- wandb/run-20240626_071628-smnm2aje/files/conda-environment.yaml +0 -0
- wandb/run-20240626_071628-smnm2aje/files/config.yaml +724 -0
- wandb/run-20240626_071628-smnm2aje/files/output.log +105 -0
- wandb/run-20240626_071628-smnm2aje/files/requirements.txt +876 -0
- wandb/run-20240626_071628-smnm2aje/files/wandb-metadata.json +66 -0
- wandb/run-20240626_071628-smnm2aje/files/wandb-summary.json +1 -0
- wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log +0 -0
- wandb/run-20240626_071628-smnm2aje/logs/debug.log +246 -0
- wandb/run-20240626_071628-smnm2aje/run-smnm2aje.wandb +0 -0
README.md
ADDED
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
library_name: peft
|
4 |
+
tags:
|
5 |
+
- generated_from_trainer
|
6 |
+
base_model: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
|
7 |
+
model-index:
|
8 |
+
- name: working
|
9 |
+
results: []
|
10 |
+
---
|
11 |
+
|
12 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
+
should probably proofread and complete it, then remove this comment. -->
|
14 |
+
|
15 |
+
# working
|
16 |
+
|
17 |
+
This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
|
18 |
+
It achieves the following results on the evaluation set:
|
19 |
+
- Loss: 1.2984
|
20 |
+
|
21 |
+
## Model description
|
22 |
+
|
23 |
+
More information needed
|
24 |
+
|
25 |
+
## Intended uses & limitations
|
26 |
+
|
27 |
+
More information needed
|
28 |
+
|
29 |
+
## Training and evaluation data
|
30 |
+
|
31 |
+
More information needed
|
32 |
+
|
33 |
+
## Training procedure
|
34 |
+
|
35 |
+
### Training hyperparameters
|
36 |
+
|
37 |
+
The following hyperparameters were used during training:
|
38 |
+
- learning_rate: 0.0002
|
39 |
+
- train_batch_size: 4
|
40 |
+
- eval_batch_size: 4
|
41 |
+
- seed: 42
|
42 |
+
- gradient_accumulation_steps: 4
|
43 |
+
- total_train_batch_size: 16
|
44 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
45 |
+
- lr_scheduler_type: linear
|
46 |
+
- lr_scheduler_warmup_steps: 2
|
47 |
+
- num_epochs: 10
|
48 |
+
- mixed_precision_training: Native AMP
|
49 |
+
|
50 |
+
### Training results
|
51 |
+
|
52 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
53 |
+
|:-------------:|:------:|:----:|:---------------:|
|
54 |
+
| 3.1139 | 0.9412 | 12 | 2.3084 |
|
55 |
+
| 1.9191 | 1.9608 | 25 | 1.7055 |
|
56 |
+
| 1.6083 | 2.9804 | 38 | 1.5741 |
|
57 |
+
| 1.496 | 4.0 | 51 | 1.4695 |
|
58 |
+
| 1.5309 | 4.9412 | 63 | 1.4038 |
|
59 |
+
| 1.3627 | 5.9608 | 76 | 1.3573 |
|
60 |
+
| 1.3111 | 6.9804 | 89 | 1.3281 |
|
61 |
+
| 1.2995 | 8.0 | 102 | 1.3092 |
|
62 |
+
| 1.3866 | 8.9412 | 114 | 1.2998 |
|
63 |
+
| 1.18 | 9.4118 | 120 | 1.2984 |
|
64 |
+
|
65 |
+
|
66 |
+
### Framework versions
|
67 |
+
|
68 |
+
- PEFT 0.11.1
|
69 |
+
- Transformers 4.41.2
|
70 |
+
- Pytorch 2.1.2
|
71 |
+
- Datasets 2.19.2
|
72 |
+
- Tokenizers 0.19.1
|
adapter_config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"alpha_pattern": {},
|
3 |
+
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": null,
|
5 |
+
"bias": "none",
|
6 |
+
"fan_in_fan_out": false,
|
7 |
+
"inference_mode": true,
|
8 |
+
"init_lora_weights": true,
|
9 |
+
"layer_replication": null,
|
10 |
+
"layers_pattern": null,
|
11 |
+
"layers_to_transform": null,
|
12 |
+
"loftq_config": {},
|
13 |
+
"lora_alpha": 32,
|
14 |
+
"lora_dropout": 0.05,
|
15 |
+
"megatron_config": null,
|
16 |
+
"megatron_core": "megatron.core",
|
17 |
+
"modules_to_save": null,
|
18 |
+
"peft_type": "LORA",
|
19 |
+
"r": 8,
|
20 |
+
"rank_pattern": {},
|
21 |
+
"revision": null,
|
22 |
+
"target_modules": [
|
23 |
+
"q_proj"
|
24 |
+
],
|
25 |
+
"task_type": "CAUSAL_LM",
|
26 |
+
"use_dora": false,
|
27 |
+
"use_rslora": false
|
28 |
+
}
|
adapter_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cf94b36f046daba8d3cb473ba0348657a9ab8c56dbc2fb787002a64c357e7fa
|
3 |
+
size 8401408
|
runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385345.adc95cf38b20.34.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bca8edc01aa31a9e70f9815dbca96d3ac87c4247cc92a4349a8cb745bdc575e3
|
3 |
+
size 5481
|
runs/Jun26_06-58-21_adc95cf38b20/events.out.tfevents.1719385423.adc95cf38b20.34.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b0060337b7a08396a73032f8dc408d21d457d6d92945573baaee52a0267a943
|
3 |
+
size 5481
|
runs/Jun26_07-07-26_adc95cf38b20/events.out.tfevents.1719385648.adc95cf38b20.34.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcba8cc9427be248c5ef089831420e28a59559d37e4d7b1fc100ffbe95dd1d22
|
3 |
+
size 16267
|
runs/Jun26_07-16-15_adc95cf38b20/events.out.tfevents.1719386177.adc95cf38b20.34.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71bffc05745c6f8799b41a6de611947dd2d1daf0eef79e856f120996c9f55035
|
3 |
+
size 5481
|
runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719386347.adc95cf38b20.34.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e7083d5109e346d61a1145b6a1fdbd73bcf0549de48bd847c6788e0072883c2
|
3 |
+
size 10559
|
runs/Jun26_07-19-05_adc95cf38b20/events.out.tfevents.1719391099.adc95cf38b20.34.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1664a1d1de74046bcd26b750d0bb5b1b704fb72256c794a139ae46f340dc4c85
|
3 |
+
size 354
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be11ac523aaa8727f2185f783bf03400ad8ca8ff6208855f6a032bd7b8a919b3
|
3 |
+
size 5112
|
wandb/debug-internal.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
wandb/debug.log
ADDED
@@ -0,0 +1,246 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
|
2 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
|
3 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
4 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
|
5 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
|
6 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
7 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
|
8 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
9 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
10 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
11 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
12 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
|
13 |
+
2024-06-26 07:16:28,967 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
14 |
+
Traceback (most recent call last):
|
15 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
16 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
17 |
+
File "/tmp/ipykernel_34/2014566126.py", line 10, in <module>
|
18 |
+
trainer.train()
|
19 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
20 |
+
return inner_training_loop(
|
21 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
22 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
23 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
24 |
+
return self.call_event("on_train_begin", args, state, control)
|
25 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
26 |
+
result = getattr(callback, event)(
|
27 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
28 |
+
self.setup(args, state, model, **kwargs)
|
29 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
30 |
+
self._wandb.init(
|
31 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
32 |
+
wandb._sentry.reraise(e)
|
33 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
34 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
35 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
36 |
+
wi.setup(kwargs)
|
37 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
38 |
+
wandb_login._login(
|
39 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
40 |
+
wlogin.prompt_api_key()
|
41 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
42 |
+
key, status = self._prompt_api_key()
|
43 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
44 |
+
key = apikey.prompt_api_key(
|
45 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
46 |
+
key = input_callback(api_ask).strip()
|
47 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
48 |
+
value = prompt_func(prompt)
|
49 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
50 |
+
raise Abort() from None
|
51 |
+
click.exceptions.Abort
|
52 |
+
2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
53 |
+
Traceback (most recent call last):
|
54 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
55 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
56 |
+
File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
|
57 |
+
trainer.train()
|
58 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
59 |
+
return inner_training_loop(
|
60 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
61 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
62 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
63 |
+
return self.call_event("on_train_begin", args, state, control)
|
64 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
65 |
+
result = getattr(callback, event)(
|
66 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
67 |
+
self.setup(args, state, model, **kwargs)
|
68 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
69 |
+
self._wandb.init(
|
70 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
71 |
+
wandb._sentry.reraise(e)
|
72 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
73 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
74 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
75 |
+
wi.setup(kwargs)
|
76 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
77 |
+
wandb_login._login(
|
78 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
79 |
+
wlogin.prompt_api_key()
|
80 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
81 |
+
key, status = self._prompt_api_key()
|
82 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
83 |
+
key = apikey.prompt_api_key(
|
84 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
85 |
+
key = input_callback(api_ask).strip()
|
86 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
87 |
+
value = prompt_func(prompt)
|
88 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
89 |
+
raise Abort() from None
|
90 |
+
click.exceptions.Abort
|
91 |
+
2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
92 |
+
Traceback (most recent call last):
|
93 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
94 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
95 |
+
File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
|
96 |
+
trainer.train()
|
97 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
98 |
+
return inner_training_loop(
|
99 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
100 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
101 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
102 |
+
return self.call_event("on_train_begin", args, state, control)
|
103 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
104 |
+
result = getattr(callback, event)(
|
105 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
106 |
+
self.setup(args, state, model, **kwargs)
|
107 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
108 |
+
self._wandb.init(
|
109 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
110 |
+
wandb._sentry.reraise(e)
|
111 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
112 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
113 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
114 |
+
wi.setup(kwargs)
|
115 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
116 |
+
wandb_login._login(
|
117 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
118 |
+
wlogin.prompt_api_key()
|
119 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
120 |
+
key, status = self._prompt_api_key()
|
121 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
122 |
+
key = apikey.prompt_api_key(
|
123 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
124 |
+
key = input_callback(api_ask).strip()
|
125 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
126 |
+
value = prompt_func(prompt)
|
127 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
128 |
+
raise Abort() from None
|
129 |
+
click.exceptions.Abort
|
130 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug.log
|
131 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log
|
132 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78720aaa5b40>
|
133 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
|
134 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
|
135 |
+
config: {}
|
136 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():610] starting backend
|
137 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
|
138 |
+
2024-06-26 07:16:28,971 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
139 |
+
2024-06-26 07:16:28,977 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
|
140 |
+
2024-06-26 07:16:28,989 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
|
141 |
+
2024-06-26 07:16:29,303 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
|
142 |
+
2024-06-26 07:16:29,306 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
|
143 |
+
2024-06-26 07:16:29,429 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
|
144 |
+
2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
145 |
+
|
146 |
+
2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
|
147 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
|
148 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
|
149 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
|
150 |
+
2024-06-26 07:16:45,663 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
|
151 |
+
2024-06-26 07:16:45,672 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
|
152 |
+
2024-06-26 07:16:45,678 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-16-15_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
|
153 |
+
2024-06-26 07:18:26,851 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
154 |
+
2024-06-26 07:18:26,851 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
155 |
+
2024-06-26 07:18:50,892 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
156 |
+
2024-06-26 07:18:50,915 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
157 |
+
2024-06-26 07:18:50,915 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
158 |
+
2024-06-26 07:18:51,909 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
159 |
+
2024-06-26 07:18:51,913 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
160 |
+
2024-06-26 07:18:51,913 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
161 |
+
2024-06-26 07:18:52,853 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
162 |
+
2024-06-26 07:18:52,855 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
163 |
+
2024-06-26 07:18:52,855 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
164 |
+
2024-06-26 07:18:53,667 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
165 |
+
2024-06-26 07:18:53,670 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
166 |
+
2024-06-26 07:18:53,670 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
167 |
+
2024-06-26 07:18:54,691 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
168 |
+
2024-06-26 07:18:54,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
169 |
+
2024-06-26 07:18:54,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
170 |
+
2024-06-26 07:18:55,617 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
171 |
+
2024-06-26 07:18:55,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
172 |
+
2024-06-26 07:18:55,682 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
173 |
+
2024-06-26 07:18:56,512 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
174 |
+
2024-06-26 07:18:56,513 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
175 |
+
2024-06-26 07:18:56,513 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
176 |
+
2024-06-26 07:18:57,999 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
177 |
+
2024-06-26 07:18:58,001 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
178 |
+
2024-06-26 07:18:58,001 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
179 |
+
2024-06-26 07:18:58,997 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
180 |
+
2024-06-26 07:18:59,009 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
181 |
+
2024-06-26 07:18:59,009 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
182 |
+
2024-06-26 07:19:00,090 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
183 |
+
2024-06-26 07:19:00,092 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
184 |
+
2024-06-26 07:19:00,092 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
185 |
+
2024-06-26 07:19:01,722 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
186 |
+
2024-06-26 07:19:01,723 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
187 |
+
2024-06-26 07:19:01,723 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
188 |
+
2024-06-26 07:19:03,098 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
189 |
+
2024-06-26 07:19:03,117 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
190 |
+
2024-06-26 07:19:03,117 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
191 |
+
2024-06-26 07:19:04,117 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
192 |
+
2024-06-26 07:19:04,212 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
193 |
+
2024-06-26 07:19:04,212 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
194 |
+
2024-06-26 07:19:05,065 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
195 |
+
2024-06-26 07:19:05,097 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
196 |
+
2024-06-26 07:19:05,097 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
197 |
+
2024-06-26 07:19:06,209 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
198 |
+
2024-06-26 07:19:06,220 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
199 |
+
2024-06-26 07:19:06,220 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
200 |
+
2024-06-26 07:19:07,300 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
201 |
+
2024-06-26 07:19:07,961 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-19-05_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
|
202 |
+
2024-06-26 08:35:48,386 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
203 |
+
2024-06-26 08:35:48,387 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
204 |
+
2024-06-26 08:36:11,677 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
205 |
+
2024-06-26 08:38:19,575 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
206 |
+
2024-06-26 08:38:19,575 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
207 |
+
2024-06-26 08:39:55,100 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
208 |
+
2024-06-26 08:39:55,103 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
209 |
+
2024-06-26 08:39:55,103 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
210 |
+
2024-06-26 08:40:32,017 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
211 |
+
2024-06-26 08:43:02,593 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
212 |
+
2024-06-26 08:43:02,594 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
213 |
+
2024-06-26 08:45:10,959 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
214 |
+
2024-06-26 08:45:10,963 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
215 |
+
2024-06-26 08:45:10,963 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
216 |
+
2024-06-26 08:45:13,384 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
217 |
+
2024-06-26 08:47:43,994 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
218 |
+
2024-06-26 08:47:43,994 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
219 |
+
2024-06-26 08:49:58,760 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
220 |
+
2024-06-26 08:53:33,771 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
221 |
+
2024-06-26 08:53:33,771 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
222 |
+
2024-06-26 08:54:32,055 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
223 |
+
2024-06-26 08:54:32,059 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
224 |
+
2024-06-26 08:54:32,059 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
225 |
+
2024-06-26 08:54:32,754 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
226 |
+
2024-06-26 08:58:07,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
227 |
+
2024-06-26 08:58:07,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
228 |
+
2024-06-26 09:12:46,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
229 |
+
2024-06-26 09:12:46,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
230 |
+
2024-06-26 09:12:46,669 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
231 |
+
2024-06-26 09:16:22,037 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
232 |
+
2024-06-26 09:16:22,038 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
233 |
+
2024-06-26 09:16:22,038 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
234 |
+
2024-06-26 09:16:35,228 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
235 |
+
2024-06-26 09:16:36,438 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
236 |
+
2024-06-26 09:16:36,438 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
237 |
+
2024-06-26 09:18:25,016 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
238 |
+
2024-06-26 09:18:26,213 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
239 |
+
2024-06-26 09:18:26,214 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
240 |
+
2024-06-26 09:21:21,819 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
241 |
+
2024-06-26 09:21:21,842 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
242 |
+
2024-06-26 09:21:21,842 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
243 |
+
2024-06-26 09:21:36,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
244 |
+
2024-06-26 09:21:36,633 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
245 |
+
2024-06-26 09:21:36,634 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
246 |
+
2024-06-26 09:21:37,142 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
wandb/run-20240626_071628-smnm2aje/files/conda-environment.yaml
ADDED
File without changes
|
wandb/run-20240626_071628-smnm2aje/files/config.yaml
ADDED
@@ -0,0 +1,724 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
wandb_version: 1
|
2 |
+
|
3 |
+
_wandb:
|
4 |
+
desc: null
|
5 |
+
value:
|
6 |
+
python_version: 3.10.13
|
7 |
+
cli_version: 0.17.0
|
8 |
+
framework: huggingface
|
9 |
+
huggingface_version: 4.41.2
|
10 |
+
is_jupyter_run: true
|
11 |
+
is_kaggle_kernel: true
|
12 |
+
start_time: 1719386188
|
13 |
+
t:
|
14 |
+
1:
|
15 |
+
- 1
|
16 |
+
- 2
|
17 |
+
- 3
|
18 |
+
- 5
|
19 |
+
- 11
|
20 |
+
- 12
|
21 |
+
- 49
|
22 |
+
- 51
|
23 |
+
- 53
|
24 |
+
- 55
|
25 |
+
- 71
|
26 |
+
- 98
|
27 |
+
- 99
|
28 |
+
- 105
|
29 |
+
2:
|
30 |
+
- 1
|
31 |
+
- 2
|
32 |
+
- 3
|
33 |
+
- 5
|
34 |
+
- 11
|
35 |
+
- 12
|
36 |
+
- 49
|
37 |
+
- 51
|
38 |
+
- 53
|
39 |
+
- 55
|
40 |
+
- 71
|
41 |
+
- 98
|
42 |
+
- 99
|
43 |
+
- 105
|
44 |
+
3:
|
45 |
+
- 7
|
46 |
+
- 13
|
47 |
+
- 23
|
48 |
+
- 62
|
49 |
+
- 66
|
50 |
+
4: 3.10.13
|
51 |
+
5: 0.17.0
|
52 |
+
6: 4.41.2
|
53 |
+
8:
|
54 |
+
- 1
|
55 |
+
- 2
|
56 |
+
- 5
|
57 |
+
9:
|
58 |
+
1: transformers_trainer
|
59 |
+
13: linux-x86_64
|
60 |
+
m:
|
61 |
+
- 1: train/global_step
|
62 |
+
6:
|
63 |
+
- 3
|
64 |
+
- 1: train/loss
|
65 |
+
5: 1
|
66 |
+
6:
|
67 |
+
- 1
|
68 |
+
- 1: train/grad_norm
|
69 |
+
5: 1
|
70 |
+
6:
|
71 |
+
- 1
|
72 |
+
- 1: train/learning_rate
|
73 |
+
5: 1
|
74 |
+
6:
|
75 |
+
- 1
|
76 |
+
- 1: train/epoch
|
77 |
+
5: 1
|
78 |
+
6:
|
79 |
+
- 1
|
80 |
+
- 1: eval/loss
|
81 |
+
5: 1
|
82 |
+
6:
|
83 |
+
- 1
|
84 |
+
- 1: eval/runtime
|
85 |
+
5: 1
|
86 |
+
6:
|
87 |
+
- 1
|
88 |
+
- 1: eval/samples_per_second
|
89 |
+
5: 1
|
90 |
+
6:
|
91 |
+
- 1
|
92 |
+
- 1: eval/steps_per_second
|
93 |
+
5: 1
|
94 |
+
6:
|
95 |
+
- 1
|
96 |
+
vocab_size:
|
97 |
+
desc: null
|
98 |
+
value: 32000
|
99 |
+
max_position_embeddings:
|
100 |
+
desc: null
|
101 |
+
value: 32768
|
102 |
+
hidden_size:
|
103 |
+
desc: null
|
104 |
+
value: 4096
|
105 |
+
intermediate_size:
|
106 |
+
desc: null
|
107 |
+
value: 14336
|
108 |
+
num_hidden_layers:
|
109 |
+
desc: null
|
110 |
+
value: 32
|
111 |
+
num_attention_heads:
|
112 |
+
desc: null
|
113 |
+
value: 32
|
114 |
+
sliding_window:
|
115 |
+
desc: null
|
116 |
+
value: null
|
117 |
+
num_key_value_heads:
|
118 |
+
desc: null
|
119 |
+
value: 8
|
120 |
+
hidden_act:
|
121 |
+
desc: null
|
122 |
+
value: silu
|
123 |
+
initializer_range:
|
124 |
+
desc: null
|
125 |
+
value: 0.02
|
126 |
+
rms_norm_eps:
|
127 |
+
desc: null
|
128 |
+
value: 1.0e-05
|
129 |
+
use_cache:
|
130 |
+
desc: null
|
131 |
+
value: false
|
132 |
+
rope_theta:
|
133 |
+
desc: null
|
134 |
+
value: 1000000.0
|
135 |
+
attention_dropout:
|
136 |
+
desc: null
|
137 |
+
value: 0.0
|
138 |
+
return_dict:
|
139 |
+
desc: null
|
140 |
+
value: true
|
141 |
+
output_hidden_states:
|
142 |
+
desc: null
|
143 |
+
value: false
|
144 |
+
output_attentions:
|
145 |
+
desc: null
|
146 |
+
value: false
|
147 |
+
torchscript:
|
148 |
+
desc: null
|
149 |
+
value: false
|
150 |
+
torch_dtype:
|
151 |
+
desc: null
|
152 |
+
value: bfloat16
|
153 |
+
use_bfloat16:
|
154 |
+
desc: null
|
155 |
+
value: false
|
156 |
+
tf_legacy_loss:
|
157 |
+
desc: null
|
158 |
+
value: false
|
159 |
+
pruned_heads:
|
160 |
+
desc: null
|
161 |
+
value: {}
|
162 |
+
tie_word_embeddings:
|
163 |
+
desc: null
|
164 |
+
value: false
|
165 |
+
chunk_size_feed_forward:
|
166 |
+
desc: null
|
167 |
+
value: 0
|
168 |
+
is_encoder_decoder:
|
169 |
+
desc: null
|
170 |
+
value: false
|
171 |
+
is_decoder:
|
172 |
+
desc: null
|
173 |
+
value: false
|
174 |
+
cross_attention_hidden_size:
|
175 |
+
desc: null
|
176 |
+
value: null
|
177 |
+
add_cross_attention:
|
178 |
+
desc: null
|
179 |
+
value: false
|
180 |
+
tie_encoder_decoder:
|
181 |
+
desc: null
|
182 |
+
value: false
|
183 |
+
max_length:
|
184 |
+
desc: null
|
185 |
+
value: 20
|
186 |
+
min_length:
|
187 |
+
desc: null
|
188 |
+
value: 0
|
189 |
+
do_sample:
|
190 |
+
desc: null
|
191 |
+
value: false
|
192 |
+
early_stopping:
|
193 |
+
desc: null
|
194 |
+
value: false
|
195 |
+
num_beams:
|
196 |
+
desc: null
|
197 |
+
value: 1
|
198 |
+
num_beam_groups:
|
199 |
+
desc: null
|
200 |
+
value: 1
|
201 |
+
diversity_penalty:
|
202 |
+
desc: null
|
203 |
+
value: 0.0
|
204 |
+
temperature:
|
205 |
+
desc: null
|
206 |
+
value: 1.0
|
207 |
+
top_k:
|
208 |
+
desc: null
|
209 |
+
value: 50
|
210 |
+
top_p:
|
211 |
+
desc: null
|
212 |
+
value: 1.0
|
213 |
+
typical_p:
|
214 |
+
desc: null
|
215 |
+
value: 1.0
|
216 |
+
repetition_penalty:
|
217 |
+
desc: null
|
218 |
+
value: 1.0
|
219 |
+
length_penalty:
|
220 |
+
desc: null
|
221 |
+
value: 1.0
|
222 |
+
no_repeat_ngram_size:
|
223 |
+
desc: null
|
224 |
+
value: 0
|
225 |
+
encoder_no_repeat_ngram_size:
|
226 |
+
desc: null
|
227 |
+
value: 0
|
228 |
+
bad_words_ids:
|
229 |
+
desc: null
|
230 |
+
value: null
|
231 |
+
num_return_sequences:
|
232 |
+
desc: null
|
233 |
+
value: 1
|
234 |
+
output_scores:
|
235 |
+
desc: null
|
236 |
+
value: false
|
237 |
+
return_dict_in_generate:
|
238 |
+
desc: null
|
239 |
+
value: false
|
240 |
+
forced_bos_token_id:
|
241 |
+
desc: null
|
242 |
+
value: null
|
243 |
+
forced_eos_token_id:
|
244 |
+
desc: null
|
245 |
+
value: null
|
246 |
+
remove_invalid_values:
|
247 |
+
desc: null
|
248 |
+
value: false
|
249 |
+
exponential_decay_length_penalty:
|
250 |
+
desc: null
|
251 |
+
value: null
|
252 |
+
suppress_tokens:
|
253 |
+
desc: null
|
254 |
+
value: null
|
255 |
+
begin_suppress_tokens:
|
256 |
+
desc: null
|
257 |
+
value: null
|
258 |
+
architectures:
|
259 |
+
desc: null
|
260 |
+
value:
|
261 |
+
- MistralForCausalLM
|
262 |
+
finetuning_task:
|
263 |
+
desc: null
|
264 |
+
value: null
|
265 |
+
id2label:
|
266 |
+
desc: null
|
267 |
+
value:
|
268 |
+
'0': LABEL_0
|
269 |
+
'1': LABEL_1
|
270 |
+
label2id:
|
271 |
+
desc: null
|
272 |
+
value:
|
273 |
+
LABEL_0: 0
|
274 |
+
LABEL_1: 1
|
275 |
+
tokenizer_class:
|
276 |
+
desc: null
|
277 |
+
value: null
|
278 |
+
prefix:
|
279 |
+
desc: null
|
280 |
+
value: null
|
281 |
+
bos_token_id:
|
282 |
+
desc: null
|
283 |
+
value: 1
|
284 |
+
pad_token_id:
|
285 |
+
desc: null
|
286 |
+
value: 0
|
287 |
+
eos_token_id:
|
288 |
+
desc: null
|
289 |
+
value: 2
|
290 |
+
sep_token_id:
|
291 |
+
desc: null
|
292 |
+
value: null
|
293 |
+
decoder_start_token_id:
|
294 |
+
desc: null
|
295 |
+
value: null
|
296 |
+
task_specific_params:
|
297 |
+
desc: null
|
298 |
+
value: null
|
299 |
+
problem_type:
|
300 |
+
desc: null
|
301 |
+
value: null
|
302 |
+
_name_or_path:
|
303 |
+
desc: null
|
304 |
+
value: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
|
305 |
+
transformers_version:
|
306 |
+
desc: null
|
307 |
+
value: 4.41.2
|
308 |
+
model_type:
|
309 |
+
desc: null
|
310 |
+
value: mistral
|
311 |
+
pretraining_tp:
|
312 |
+
desc: null
|
313 |
+
value: 1
|
314 |
+
quantization_config:
|
315 |
+
desc: null
|
316 |
+
value:
|
317 |
+
quant_method: QuantizationMethod.GPTQ
|
318 |
+
bits: 4
|
319 |
+
tokenizer: null
|
320 |
+
dataset: null
|
321 |
+
group_size: 128
|
322 |
+
damp_percent: 0.1
|
323 |
+
desc_act: true
|
324 |
+
sym: true
|
325 |
+
true_sequential: true
|
326 |
+
use_cuda_fp16: false
|
327 |
+
model_seqlen: null
|
328 |
+
block_name_to_quantize: null
|
329 |
+
module_name_preceding_first_block: null
|
330 |
+
batch_size: 1
|
331 |
+
pad_token_id: null
|
332 |
+
use_exllama: true
|
333 |
+
max_input_length: null
|
334 |
+
exllama_config:
|
335 |
+
version: ExllamaVersion.ONE
|
336 |
+
cache_block_outputs: true
|
337 |
+
modules_in_block_to_quantize: null
|
338 |
+
output_dir:
|
339 |
+
desc: null
|
340 |
+
value: /kaggle/working/
|
341 |
+
overwrite_output_dir:
|
342 |
+
desc: null
|
343 |
+
value: false
|
344 |
+
do_train:
|
345 |
+
desc: null
|
346 |
+
value: false
|
347 |
+
do_eval:
|
348 |
+
desc: null
|
349 |
+
value: true
|
350 |
+
do_predict:
|
351 |
+
desc: null
|
352 |
+
value: false
|
353 |
+
eval_strategy:
|
354 |
+
desc: null
|
355 |
+
value: epoch
|
356 |
+
prediction_loss_only:
|
357 |
+
desc: null
|
358 |
+
value: false
|
359 |
+
per_device_train_batch_size:
|
360 |
+
desc: null
|
361 |
+
value: 4
|
362 |
+
per_device_eval_batch_size:
|
363 |
+
desc: null
|
364 |
+
value: 4
|
365 |
+
per_gpu_train_batch_size:
|
366 |
+
desc: null
|
367 |
+
value: null
|
368 |
+
per_gpu_eval_batch_size:
|
369 |
+
desc: null
|
370 |
+
value: null
|
371 |
+
gradient_accumulation_steps:
|
372 |
+
desc: null
|
373 |
+
value: 4
|
374 |
+
eval_accumulation_steps:
|
375 |
+
desc: null
|
376 |
+
value: null
|
377 |
+
eval_delay:
|
378 |
+
desc: null
|
379 |
+
value: 0
|
380 |
+
learning_rate:
|
381 |
+
desc: null
|
382 |
+
value: 0.0002
|
383 |
+
weight_decay:
|
384 |
+
desc: null
|
385 |
+
value: 0.01
|
386 |
+
adam_beta1:
|
387 |
+
desc: null
|
388 |
+
value: 0.9
|
389 |
+
adam_beta2:
|
390 |
+
desc: null
|
391 |
+
value: 0.999
|
392 |
+
adam_epsilon:
|
393 |
+
desc: null
|
394 |
+
value: 1.0e-08
|
395 |
+
max_grad_norm:
|
396 |
+
desc: null
|
397 |
+
value: 1.0
|
398 |
+
num_train_epochs:
|
399 |
+
desc: null
|
400 |
+
value: 10
|
401 |
+
max_steps:
|
402 |
+
desc: null
|
403 |
+
value: -1
|
404 |
+
lr_scheduler_type:
|
405 |
+
desc: null
|
406 |
+
value: linear
|
407 |
+
lr_scheduler_kwargs:
|
408 |
+
desc: null
|
409 |
+
value: {}
|
410 |
+
warmup_ratio:
|
411 |
+
desc: null
|
412 |
+
value: 0.0
|
413 |
+
warmup_steps:
|
414 |
+
desc: null
|
415 |
+
value: 2
|
416 |
+
log_level:
|
417 |
+
desc: null
|
418 |
+
value: passive
|
419 |
+
log_level_replica:
|
420 |
+
desc: null
|
421 |
+
value: warning
|
422 |
+
log_on_each_node:
|
423 |
+
desc: null
|
424 |
+
value: true
|
425 |
+
logging_dir:
|
426 |
+
desc: null
|
427 |
+
value: /kaggle/working/runs/Jun26_07-19-05_adc95cf38b20
|
428 |
+
logging_strategy:
|
429 |
+
desc: null
|
430 |
+
value: epoch
|
431 |
+
logging_first_step:
|
432 |
+
desc: null
|
433 |
+
value: false
|
434 |
+
logging_steps:
|
435 |
+
desc: null
|
436 |
+
value: 500
|
437 |
+
logging_nan_inf_filter:
|
438 |
+
desc: null
|
439 |
+
value: true
|
440 |
+
save_strategy:
|
441 |
+
desc: null
|
442 |
+
value: epoch
|
443 |
+
save_steps:
|
444 |
+
desc: null
|
445 |
+
value: 500
|
446 |
+
save_total_limit:
|
447 |
+
desc: null
|
448 |
+
value: null
|
449 |
+
save_safetensors:
|
450 |
+
desc: null
|
451 |
+
value: true
|
452 |
+
save_on_each_node:
|
453 |
+
desc: null
|
454 |
+
value: false
|
455 |
+
save_only_model:
|
456 |
+
desc: null
|
457 |
+
value: false
|
458 |
+
restore_callback_states_from_checkpoint:
|
459 |
+
desc: null
|
460 |
+
value: false
|
461 |
+
no_cuda:
|
462 |
+
desc: null
|
463 |
+
value: false
|
464 |
+
use_cpu:
|
465 |
+
desc: null
|
466 |
+
value: false
|
467 |
+
use_mps_device:
|
468 |
+
desc: null
|
469 |
+
value: false
|
470 |
+
seed:
|
471 |
+
desc: null
|
472 |
+
value: 42
|
473 |
+
data_seed:
|
474 |
+
desc: null
|
475 |
+
value: null
|
476 |
+
jit_mode_eval:
|
477 |
+
desc: null
|
478 |
+
value: false
|
479 |
+
use_ipex:
|
480 |
+
desc: null
|
481 |
+
value: false
|
482 |
+
bf16:
|
483 |
+
desc: null
|
484 |
+
value: false
|
485 |
+
fp16:
|
486 |
+
desc: null
|
487 |
+
value: true
|
488 |
+
fp16_opt_level:
|
489 |
+
desc: null
|
490 |
+
value: O1
|
491 |
+
half_precision_backend:
|
492 |
+
desc: null
|
493 |
+
value: auto
|
494 |
+
bf16_full_eval:
|
495 |
+
desc: null
|
496 |
+
value: false
|
497 |
+
fp16_full_eval:
|
498 |
+
desc: null
|
499 |
+
value: false
|
500 |
+
tf32:
|
501 |
+
desc: null
|
502 |
+
value: null
|
503 |
+
local_rank:
|
504 |
+
desc: null
|
505 |
+
value: 0
|
506 |
+
ddp_backend:
|
507 |
+
desc: null
|
508 |
+
value: null
|
509 |
+
tpu_num_cores:
|
510 |
+
desc: null
|
511 |
+
value: null
|
512 |
+
tpu_metrics_debug:
|
513 |
+
desc: null
|
514 |
+
value: false
|
515 |
+
debug:
|
516 |
+
desc: null
|
517 |
+
value: []
|
518 |
+
dataloader_drop_last:
|
519 |
+
desc: null
|
520 |
+
value: false
|
521 |
+
eval_steps:
|
522 |
+
desc: null
|
523 |
+
value: null
|
524 |
+
dataloader_num_workers:
|
525 |
+
desc: null
|
526 |
+
value: 0
|
527 |
+
dataloader_prefetch_factor:
|
528 |
+
desc: null
|
529 |
+
value: null
|
530 |
+
past_index:
|
531 |
+
desc: null
|
532 |
+
value: -1
|
533 |
+
run_name:
|
534 |
+
desc: null
|
535 |
+
value: /kaggle/working/
|
536 |
+
disable_tqdm:
|
537 |
+
desc: null
|
538 |
+
value: false
|
539 |
+
remove_unused_columns:
|
540 |
+
desc: null
|
541 |
+
value: true
|
542 |
+
label_names:
|
543 |
+
desc: null
|
544 |
+
value: null
|
545 |
+
load_best_model_at_end:
|
546 |
+
desc: null
|
547 |
+
value: true
|
548 |
+
metric_for_best_model:
|
549 |
+
desc: null
|
550 |
+
value: loss
|
551 |
+
greater_is_better:
|
552 |
+
desc: null
|
553 |
+
value: false
|
554 |
+
ignore_data_skip:
|
555 |
+
desc: null
|
556 |
+
value: false
|
557 |
+
fsdp:
|
558 |
+
desc: null
|
559 |
+
value: []
|
560 |
+
fsdp_min_num_params:
|
561 |
+
desc: null
|
562 |
+
value: 0
|
563 |
+
fsdp_config:
|
564 |
+
desc: null
|
565 |
+
value:
|
566 |
+
min_num_params: 0
|
567 |
+
xla: false
|
568 |
+
xla_fsdp_v2: false
|
569 |
+
xla_fsdp_grad_ckpt: false
|
570 |
+
fsdp_transformer_layer_cls_to_wrap:
|
571 |
+
desc: null
|
572 |
+
value: null
|
573 |
+
accelerator_config:
|
574 |
+
desc: null
|
575 |
+
value:
|
576 |
+
split_batches: false
|
577 |
+
dispatch_batches: null
|
578 |
+
even_batches: true
|
579 |
+
use_seedable_sampler: true
|
580 |
+
non_blocking: false
|
581 |
+
gradient_accumulation_kwargs: null
|
582 |
+
deepspeed:
|
583 |
+
desc: null
|
584 |
+
value: null
|
585 |
+
label_smoothing_factor:
|
586 |
+
desc: null
|
587 |
+
value: 0.0
|
588 |
+
optim:
|
589 |
+
desc: null
|
590 |
+
value: paged_adamw_8bit
|
591 |
+
optim_args:
|
592 |
+
desc: null
|
593 |
+
value: null
|
594 |
+
adafactor:
|
595 |
+
desc: null
|
596 |
+
value: false
|
597 |
+
group_by_length:
|
598 |
+
desc: null
|
599 |
+
value: false
|
600 |
+
length_column_name:
|
601 |
+
desc: null
|
602 |
+
value: length
|
603 |
+
report_to:
|
604 |
+
desc: null
|
605 |
+
value:
|
606 |
+
- tensorboard
|
607 |
+
- wandb
|
608 |
+
ddp_find_unused_parameters:
|
609 |
+
desc: null
|
610 |
+
value: null
|
611 |
+
ddp_bucket_cap_mb:
|
612 |
+
desc: null
|
613 |
+
value: null
|
614 |
+
ddp_broadcast_buffers:
|
615 |
+
desc: null
|
616 |
+
value: null
|
617 |
+
dataloader_pin_memory:
|
618 |
+
desc: null
|
619 |
+
value: true
|
620 |
+
dataloader_persistent_workers:
|
621 |
+
desc: null
|
622 |
+
value: false
|
623 |
+
skip_memory_metrics:
|
624 |
+
desc: null
|
625 |
+
value: true
|
626 |
+
use_legacy_prediction_loop:
|
627 |
+
desc: null
|
628 |
+
value: false
|
629 |
+
push_to_hub:
|
630 |
+
desc: null
|
631 |
+
value: false
|
632 |
+
resume_from_checkpoint:
|
633 |
+
desc: null
|
634 |
+
value: null
|
635 |
+
hub_model_id:
|
636 |
+
desc: null
|
637 |
+
value: null
|
638 |
+
hub_strategy:
|
639 |
+
desc: null
|
640 |
+
value: every_save
|
641 |
+
hub_token:
|
642 |
+
desc: null
|
643 |
+
value: <HUB_TOKEN>
|
644 |
+
hub_private_repo:
|
645 |
+
desc: null
|
646 |
+
value: false
|
647 |
+
hub_always_push:
|
648 |
+
desc: null
|
649 |
+
value: false
|
650 |
+
gradient_checkpointing:
|
651 |
+
desc: null
|
652 |
+
value: false
|
653 |
+
gradient_checkpointing_kwargs:
|
654 |
+
desc: null
|
655 |
+
value: null
|
656 |
+
include_inputs_for_metrics:
|
657 |
+
desc: null
|
658 |
+
value: false
|
659 |
+
eval_do_concat_batches:
|
660 |
+
desc: null
|
661 |
+
value: true
|
662 |
+
fp16_backend:
|
663 |
+
desc: null
|
664 |
+
value: auto
|
665 |
+
evaluation_strategy:
|
666 |
+
desc: null
|
667 |
+
value: null
|
668 |
+
push_to_hub_model_id:
|
669 |
+
desc: null
|
670 |
+
value: null
|
671 |
+
push_to_hub_organization:
|
672 |
+
desc: null
|
673 |
+
value: null
|
674 |
+
push_to_hub_token:
|
675 |
+
desc: null
|
676 |
+
value: <PUSH_TO_HUB_TOKEN>
|
677 |
+
mp_parameters:
|
678 |
+
desc: null
|
679 |
+
value: ''
|
680 |
+
auto_find_batch_size:
|
681 |
+
desc: null
|
682 |
+
value: false
|
683 |
+
full_determinism:
|
684 |
+
desc: null
|
685 |
+
value: false
|
686 |
+
torchdynamo:
|
687 |
+
desc: null
|
688 |
+
value: null
|
689 |
+
ray_scope:
|
690 |
+
desc: null
|
691 |
+
value: last
|
692 |
+
ddp_timeout:
|
693 |
+
desc: null
|
694 |
+
value: 1800
|
695 |
+
torch_compile:
|
696 |
+
desc: null
|
697 |
+
value: false
|
698 |
+
torch_compile_backend:
|
699 |
+
desc: null
|
700 |
+
value: null
|
701 |
+
torch_compile_mode:
|
702 |
+
desc: null
|
703 |
+
value: null
|
704 |
+
dispatch_batches:
|
705 |
+
desc: null
|
706 |
+
value: null
|
707 |
+
split_batches:
|
708 |
+
desc: null
|
709 |
+
value: null
|
710 |
+
include_tokens_per_second:
|
711 |
+
desc: null
|
712 |
+
value: false
|
713 |
+
include_num_input_tokens_seen:
|
714 |
+
desc: null
|
715 |
+
value: false
|
716 |
+
neftune_noise_alpha:
|
717 |
+
desc: null
|
718 |
+
value: null
|
719 |
+
optim_target_modules:
|
720 |
+
desc: null
|
721 |
+
value: null
|
722 |
+
batch_eval_metrics:
|
723 |
+
desc: null
|
724 |
+
value: false
|
wandb/run-20240626_071628-smnm2aje/files/output.log
ADDED
@@ -0,0 +1,105 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
2 |
+
warnings.warn(
|
3 |
+
KeyboardInterrupt
|
4 |
+
<s>[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
5 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
6 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
7 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
8 |
+
please provide the valuable advice for this comment.
|
9 |
+
I'm going through some things with my feelings and myself. I barely sleep and I do nothing but think about how I'm worthless and how I shouldn't be here.
|
10 |
+
I've never tried or contemplated suicide. I've always wanted to fix my issues, but I never get around to it.
|
11 |
+
How can I change my feeling of being worthless to everyone?
|
12 |
+
[/INST]
|
13 |
+
If everyone thinks you're worthless, then maybe you need to find new people to hang out with.Seriously, the social context in which a person lives is a big influence in self-esteem.Otherwise, you can go round and round trying to understand why you're not worthless, then go back to the same crowd and be knocked down again.There are many inspirational messages you can find in social media. Maybe read some of the ones which state that no person is worthless, and that everyone has a good purpose to their life.Also, since our culture is so saturated with the belief that if someone doesn't feel good about themselves that this is somehow terrible.Bad feelings are part of living. They are the motivation to remove ourselves from situations and relationships which do us more harm than good.Bad feelings do feel terrible. Your feeling of worthlessness may be good in the sense of motivating you to find out that you are much better than your feelings today.
|
14 |
+
</s>
|
15 |
+
<s>[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
16 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
17 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
18 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
19 |
+
please provide the valuable advice for this comment.
|
20 |
+
I've never been able to talk with my parents. My parents are in their sixties while I am a teenager. I love both of them but not their personalities. I feel that they do not take me seriously whenever I talk about a serious event in my life. If my dad doesn’t believe me, then my mom goes along with my dad and acts like she doesn’t believe me either. I’m a pansexual, but I can’t trust my own parents.
|
21 |
+
I've fought depression and won; however, stress and anxiety are killing me. I feel that my friends don't listen to me. I know they have their own problems, which I do my best to help with. But they don't always try to help me with mine, when I really need them. I feel as if my childhood has been taken from me. I feel as if I have no one whom I can trust.
|
22 |
+
[/INST]
|
23 |
+
Life sounds extremely frustrating for you right now.As long as you've reached your conclusion that your parents do not take you seriously when you would like to speak with them on a serious topic, and that you feel unable to trust either of them, then work with your own information.Many people don't know how to listen and aren't trustworthy. As painful and aggravating as having parents who fit this description, you're better off to honor what you realize than to continue fighting what is not in your hands to change.Since you write about having a similar problem with friends, then maybe this comes about because you pick friends who are similarly unsatisfying to be with as are your parents.The first step out of this dilemma would be to accept how crummy you feel about the way your parents are toward you.If you're able to live with the sadness and resentment of having parents who don't want to know how to understand you, then you'll start to develop a new and more satisfying definition of what you expect from the behaviors of your friends.Trust yourself that what you feel and see is accurate, then it becomes easier to trust and welcome new people who are formatted differently than either your parents or your friends who sound very similar to your parents' ways.
|
24 |
+
</s>
|
25 |
+
trainable params: 2,097,152 || all params: 264,507,392 || trainable%: 0.7929
|
26 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
27 |
+
warnings.warn(
|
28 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
29 |
+
warnings.warn(
|
30 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
31 |
+
warnings.warn(
|
32 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
33 |
+
warnings.warn(
|
34 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
35 |
+
warnings.warn(
|
36 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
37 |
+
warnings.warn(
|
38 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
39 |
+
warnings.warn(
|
40 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
41 |
+
warnings.warn(
|
42 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
43 |
+
warnings.warn(
|
44 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
45 |
+
warnings.warn(
|
46 |
+
[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
47 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
48 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
49 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
50 |
+
please provide the valuable advice for this comment.
|
51 |
+
i think i do not fit in this world
|
52 |
+
[/INST]
|
53 |
+
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
|
54 |
+
Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
|
55 |
+
<s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
56 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
57 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
58 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
59 |
+
please provide the valuable advice for this comment.
|
60 |
+
i think i do not fit in this world
|
61 |
+
[/INST]
|
62 |
+
i think you are just trying to find your place in the world. It is okay to not fit in with everyone. It is okay to be different. It is okay to be yourself. It is okay to not be like everyone else. It is okay to be unique. It is okay to be you. It is okay to be happy. It is okay to be sad. It is okay to be angry. It is okay to be happy. It is okay to be sad. It is okay to be angry. It is okay to be happy.
|
63 |
+
[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
64 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
65 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
66 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
67 |
+
please provide the valuable advice for this comment.
|
68 |
+
I am not feeling good, my wife cheated on me , i think i should commit suicide
|
69 |
+
[/INST]
|
70 |
+
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
|
71 |
+
Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
|
72 |
+
<s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
73 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
74 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
75 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
76 |
+
please provide the valuable advice for this comment.
|
77 |
+
I am not feeling good, my wife cheated on me , i think i should commit suicide
|
78 |
+
[/INST]
|
79 |
+
I'm sorry to hear that you are feeling this way. It is important to remember that you are not alone and that there are people who want to help you. I would encourage you to reach out to a therapist or counselor who can help you work through these feelings. If you are not comfortable doing that, there are also hotlines that you can call. The National Suicide Prevention Lifeline is 1-800-273-TALK (1-800-273-8255). You can also call the Crisis Text Line by texting 7
|
80 |
+
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
|
81 |
+
Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
|
82 |
+
<s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
83 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
84 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
85 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
86 |
+
please provide the valuable advice for this comment.
|
87 |
+
I am not feeling good, my wife cheated on me , i think i should commit suicide
|
88 |
+
[/INST]
|
89 |
+
I'm sorry to hear that you are feeling this way. It is important to remember that you are not alone and that there are people who want to help you. I would encourage you to reach out to a therapist or counselor who can help you work through these feelings. If you are not comfortable doing that, there are also hotlines that you can call. The National Suicide Prevention Lifeline is 1-800-273-TALK (1-800-273-8255). You can also call the Crisis Text Line by texting 741741. I would also encourage you to consider joining a support group. Support groups can be a great resource for people who are dealing with the aftermath of an affair. You can search for a support group in your area by going to the website of
|
90 |
+
[INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
91 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
92 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
93 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
94 |
+
please provide the valuable advice for this comment.
|
95 |
+
I am not feeling good, i think life is boring
|
96 |
+
[/INST]
|
97 |
+
The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
|
98 |
+
Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
|
99 |
+
<s> [INST] You are a compassionate and knowledgeable mental health advisor. Someone is sharing their mental
|
100 |
+
state or expressing doubts about their well-being. Provide the best advice and suggestions to help them navigate their
|
101 |
+
feelings. Offer empathetic, practical, and actionable guidance, and recommend helpful resources or techniques.
|
102 |
+
Ensure your response is simple, relatable, encourages openness, and provides reassurance.
|
103 |
+
please provide the valuable advice for this comment.
|
104 |
+
I am not feeling good, i think life is boring
|
105 |
+
[/INST]
|
wandb/run-20240626_071628-smnm2aje/files/requirements.txt
ADDED
@@ -0,0 +1,876 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Babel==2.14.0
|
2 |
+
Boruta==0.3
|
3 |
+
Brotli==1.1.0
|
4 |
+
CVXcanon==0.1.2
|
5 |
+
Cartopy==0.23.0
|
6 |
+
Cython==3.0.8
|
7 |
+
Deprecated==1.2.14
|
8 |
+
Farama-Notifications==0.0.4
|
9 |
+
Flask==3.0.3
|
10 |
+
Geohash==1.0
|
11 |
+
GitPython==3.1.41
|
12 |
+
ImageHash==4.3.1
|
13 |
+
Janome==0.5.0
|
14 |
+
Jinja2==3.1.2
|
15 |
+
LunarCalendar==0.0.9
|
16 |
+
Mako==1.3.5
|
17 |
+
Markdown==3.5.2
|
18 |
+
MarkupSafe==2.1.3
|
19 |
+
MarkupSafe==2.1.5
|
20 |
+
Pillow==9.5.0
|
21 |
+
PuLP==2.8.0
|
22 |
+
PyArabic==0.6.15
|
23 |
+
PyJWT==2.8.0
|
24 |
+
PyMeeus==0.5.12
|
25 |
+
PySocks==1.7.1
|
26 |
+
PyUpSet==0.1.1.post7
|
27 |
+
PyWavelets==1.5.0
|
28 |
+
PyYAML==6.0.1
|
29 |
+
Pygments==2.17.2
|
30 |
+
Pympler==1.0.1
|
31 |
+
QtPy==2.4.1
|
32 |
+
Rtree==1.2.0
|
33 |
+
SQLAlchemy==2.0.25
|
34 |
+
SecretStorage==3.3.3
|
35 |
+
Send2Trash==1.8.2
|
36 |
+
Shapely==1.8.5.post1
|
37 |
+
Shimmy==1.3.0
|
38 |
+
SimpleITK==2.3.1
|
39 |
+
TPOT==0.12.1
|
40 |
+
Theano-PyMC==1.1.2
|
41 |
+
Theano==1.0.5
|
42 |
+
Wand==0.6.13
|
43 |
+
Werkzeug==3.0.3
|
44 |
+
absl-py==1.4.0
|
45 |
+
accelerate==0.30.1
|
46 |
+
access==1.1.9
|
47 |
+
affine==2.4.0
|
48 |
+
aiobotocore==2.13.0
|
49 |
+
aiofiles==22.1.0
|
50 |
+
aiohttp==3.9.1
|
51 |
+
aioitertools==0.11.0
|
52 |
+
aiorwlock==1.3.0
|
53 |
+
aiosignal==1.3.1
|
54 |
+
aiosqlite==0.19.0
|
55 |
+
albumentations==1.4.0
|
56 |
+
alembic==1.13.1
|
57 |
+
altair==5.3.0
|
58 |
+
annotated-types==0.6.0
|
59 |
+
annotated-types==0.7.0
|
60 |
+
annoy==1.17.3
|
61 |
+
anyio==4.2.0
|
62 |
+
apache-beam==2.46.0
|
63 |
+
aplus==0.11.0
|
64 |
+
appdirs==1.4.4
|
65 |
+
archspec==0.2.3
|
66 |
+
argon2-cffi-bindings==21.2.0
|
67 |
+
argon2-cffi==23.1.0
|
68 |
+
array-record==0.5.0
|
69 |
+
arrow==1.3.0
|
70 |
+
arviz==0.18.0
|
71 |
+
astroid==3.2.2
|
72 |
+
astropy-iers-data==0.2024.6.3.0.31.14
|
73 |
+
astropy==6.1.0
|
74 |
+
asttokens==2.4.1
|
75 |
+
astunparse==1.6.3
|
76 |
+
async-lru==2.0.4
|
77 |
+
async-timeout==4.0.3
|
78 |
+
attrs==23.2.0
|
79 |
+
audioread==3.0.1
|
80 |
+
auto_gptq==0.7.1
|
81 |
+
autopep8==2.0.4
|
82 |
+
backoff==2.2.1
|
83 |
+
bayesian-optimization==1.4.3
|
84 |
+
beatrix_jupyterlab==2023.128.151533
|
85 |
+
beautifulsoup4==4.12.2
|
86 |
+
bitsandbytes==0.43.1
|
87 |
+
blake3==0.2.1
|
88 |
+
bleach==6.1.0
|
89 |
+
blessed==1.20.0
|
90 |
+
blinker==1.8.2
|
91 |
+
blis==0.7.10
|
92 |
+
blosc2==2.6.2
|
93 |
+
bokeh==3.4.1
|
94 |
+
boltons==23.1.1
|
95 |
+
boto3==1.26.100
|
96 |
+
botocore==1.34.106
|
97 |
+
bq_helper==0.4.1
|
98 |
+
bqplot==0.12.43
|
99 |
+
branca==0.7.2
|
100 |
+
brewer2mpl==1.4.1
|
101 |
+
brotlipy==0.7.0
|
102 |
+
cached-property==1.5.2
|
103 |
+
cachetools==4.2.4
|
104 |
+
cachetools==5.3.2
|
105 |
+
catalogue==2.0.10
|
106 |
+
catalyst==22.4
|
107 |
+
catboost==1.2.5
|
108 |
+
category-encoders==2.6.3
|
109 |
+
certifi==2024.2.2
|
110 |
+
cesium==0.12.1
|
111 |
+
cffi==1.16.0
|
112 |
+
charset-normalizer==3.3.2
|
113 |
+
chex==0.1.86
|
114 |
+
cleverhans==4.0.0
|
115 |
+
click-plugins==1.1.1
|
116 |
+
click==8.1.7
|
117 |
+
cligj==0.7.2
|
118 |
+
cloud-tpu-client==0.10
|
119 |
+
cloud-tpu-profiler==2.4.0
|
120 |
+
cloudpathlib==0.16.0
|
121 |
+
cloudpickle==2.2.1
|
122 |
+
cloudpickle==3.0.0
|
123 |
+
cmdstanpy==1.2.3
|
124 |
+
colorama==0.4.6
|
125 |
+
colorcet==3.1.0
|
126 |
+
coloredlogs==15.0.1
|
127 |
+
colorful==0.5.6
|
128 |
+
colorlog==6.8.2
|
129 |
+
colorlover==0.3.0
|
130 |
+
comm==0.2.1
|
131 |
+
conda-libmamba-solver==23.12.0
|
132 |
+
conda-package-handling==2.2.0
|
133 |
+
conda==24.5.0
|
134 |
+
conda_package_streaming==0.9.0
|
135 |
+
confection==0.1.4
|
136 |
+
contextily==1.6.0
|
137 |
+
contourpy==1.2.0
|
138 |
+
contourpy==1.2.1
|
139 |
+
convertdate==2.4.0
|
140 |
+
crcmod==1.7
|
141 |
+
cryptography==41.0.7
|
142 |
+
cuda-python==12.5.0
|
143 |
+
cudf==24.4.1
|
144 |
+
cufflinks==0.17.3
|
145 |
+
cuml==24.4.0
|
146 |
+
cupy==13.1.0
|
147 |
+
cycler==0.12.1
|
148 |
+
cymem==2.0.8
|
149 |
+
cytoolz==0.12.3
|
150 |
+
daal4py==2024.4.0
|
151 |
+
daal==2024.4.0
|
152 |
+
dacite==1.8.1
|
153 |
+
dask-cuda==24.4.0
|
154 |
+
dask-cudf==24.4.1
|
155 |
+
dask-expr==1.1.2
|
156 |
+
dask==2024.5.2
|
157 |
+
dataclasses-json==0.6.6
|
158 |
+
dataproc_jupyter_plugin==0.1.66
|
159 |
+
datasets==2.19.2
|
160 |
+
datashader==0.16.2
|
161 |
+
datatile==1.0.3
|
162 |
+
db-dtypes==1.2.0
|
163 |
+
deap==1.4.1
|
164 |
+
debugpy==1.8.0
|
165 |
+
decorator==5.1.1
|
166 |
+
deepdiff==7.0.1
|
167 |
+
defusedxml==0.7.1
|
168 |
+
deprecation==2.1.0
|
169 |
+
descartes==1.1.0
|
170 |
+
dill==0.3.8
|
171 |
+
dipy==1.9.0
|
172 |
+
distlib==0.3.8
|
173 |
+
distributed==2024.1.1
|
174 |
+
distro==1.9.0
|
175 |
+
dm-tree==0.1.8
|
176 |
+
docker-pycreds==0.4.0
|
177 |
+
docker==7.0.0
|
178 |
+
docopt==0.6.2
|
179 |
+
docstring-parser==0.15
|
180 |
+
docstring-to-markdown==0.15
|
181 |
+
docutils==0.21.2
|
182 |
+
earthengine-api==0.1.405
|
183 |
+
easydict==1.13
|
184 |
+
easyocr==1.7.1
|
185 |
+
ecos==2.0.13
|
186 |
+
eli5==0.13.0
|
187 |
+
emoji==2.12.1
|
188 |
+
en-core-web-lg==3.7.1
|
189 |
+
en-core-web-sm==3.7.1
|
190 |
+
entrypoints==0.4
|
191 |
+
ephem==4.1.5
|
192 |
+
esda==2.5.1
|
193 |
+
essentia==2.1b6.dev1110
|
194 |
+
et-xmlfile==1.1.0
|
195 |
+
etils==1.6.0
|
196 |
+
exceptiongroup==1.2.0
|
197 |
+
executing==2.0.1
|
198 |
+
explainable-ai-sdk==1.3.3
|
199 |
+
fastai==2.7.15
|
200 |
+
fastapi==0.108.0
|
201 |
+
fastavro==1.9.3
|
202 |
+
fastcore==1.5.43
|
203 |
+
fastdownload==0.0.7
|
204 |
+
fasteners==0.19
|
205 |
+
fastjsonschema==2.19.1
|
206 |
+
fastprogress==1.0.3
|
207 |
+
fastrlock==0.8.2
|
208 |
+
fasttext==0.9.2
|
209 |
+
feather-format==0.4.1
|
210 |
+
featuretools==1.31.0
|
211 |
+
filelock==3.13.1
|
212 |
+
fiona==1.9.6
|
213 |
+
fitter==1.7.0
|
214 |
+
flake8==7.0.0
|
215 |
+
flashtext==2.7
|
216 |
+
flatbuffers==23.5.26
|
217 |
+
flax==0.8.4
|
218 |
+
folium==0.16.0
|
219 |
+
fonttools==4.47.0
|
220 |
+
fonttools==4.53.0
|
221 |
+
fqdn==1.5.1
|
222 |
+
frozendict==2.4.4
|
223 |
+
frozenlist==1.4.1
|
224 |
+
fsspec==2024.3.1
|
225 |
+
fsspec==2024.5.0
|
226 |
+
funcy==2.0
|
227 |
+
fury==0.10.0
|
228 |
+
future==1.0.0
|
229 |
+
fuzzywuzzy==0.18.0
|
230 |
+
gast==0.5.4
|
231 |
+
gatspy==0.3
|
232 |
+
gcsfs==2024.3.1
|
233 |
+
gekko==1.1.3
|
234 |
+
gensim==4.3.2
|
235 |
+
geographiclib==2.0
|
236 |
+
geojson==3.1.0
|
237 |
+
geopandas==0.14.4
|
238 |
+
geoplot==0.5.1
|
239 |
+
geopy==2.4.1
|
240 |
+
geoviews==1.12.0
|
241 |
+
ggplot==0.11.5
|
242 |
+
giddy==2.3.5
|
243 |
+
gitdb==4.0.11
|
244 |
+
google-ai-generativelanguage==0.6.4
|
245 |
+
google-api-core==2.11.1
|
246 |
+
google-api-core==2.19.0
|
247 |
+
google-api-python-client==2.131.0
|
248 |
+
google-apitools==0.5.31
|
249 |
+
google-auth-httplib2==0.2.0
|
250 |
+
google-auth-oauthlib==1.2.0
|
251 |
+
google-auth==2.26.1
|
252 |
+
google-cloud-aiplatform==0.6.0a1
|
253 |
+
google-cloud-artifact-registry==1.10.0
|
254 |
+
google-cloud-automl==1.0.1
|
255 |
+
google-cloud-bigquery==2.34.4
|
256 |
+
google-cloud-bigtable==1.7.3
|
257 |
+
google-cloud-core==2.4.1
|
258 |
+
google-cloud-datastore==2.19.0
|
259 |
+
google-cloud-dlp==3.14.0
|
260 |
+
google-cloud-jupyter-config==0.0.5
|
261 |
+
google-cloud-language==2.13.3
|
262 |
+
google-cloud-monitoring==2.18.0
|
263 |
+
google-cloud-pubsub==2.19.0
|
264 |
+
google-cloud-pubsublite==1.9.0
|
265 |
+
google-cloud-recommendations-ai==0.7.1
|
266 |
+
google-cloud-resource-manager==1.11.0
|
267 |
+
google-cloud-spanner==3.40.1
|
268 |
+
google-cloud-storage==1.44.0
|
269 |
+
google-cloud-translate==3.12.1
|
270 |
+
google-cloud-videointelligence==2.13.3
|
271 |
+
google-cloud-vision==2.8.0
|
272 |
+
google-crc32c==1.5.0
|
273 |
+
google-generativeai==0.6.0
|
274 |
+
google-pasta==0.2.0
|
275 |
+
google-resumable-media==2.7.0
|
276 |
+
googleapis-common-protos==1.62.0
|
277 |
+
gplearn==0.4.2
|
278 |
+
gpustat==1.0.0
|
279 |
+
gpxpy==1.6.2
|
280 |
+
graphviz==0.20.3
|
281 |
+
greenlet==3.0.3
|
282 |
+
grpc-google-iam-v1==0.12.7
|
283 |
+
grpcio-status==1.48.1
|
284 |
+
grpcio-status==1.48.2
|
285 |
+
grpcio==1.59.3
|
286 |
+
grpcio==1.60.0
|
287 |
+
gviz-api==1.10.0
|
288 |
+
gym-notices==0.0.8
|
289 |
+
gym==0.26.2
|
290 |
+
gymnasium==0.29.0
|
291 |
+
h11==0.14.0
|
292 |
+
h2o==3.46.0.2
|
293 |
+
h5netcdf==1.3.0
|
294 |
+
h5py==3.10.0
|
295 |
+
haversine==2.8.1
|
296 |
+
hdfs==2.7.3
|
297 |
+
hep-ml==0.7.2
|
298 |
+
hijri-converter==2.3.1
|
299 |
+
hmmlearn==0.3.2
|
300 |
+
holidays==0.24
|
301 |
+
holoviews==1.18.3
|
302 |
+
hpsklearn==0.1.0
|
303 |
+
html5lib==1.1
|
304 |
+
htmlmin==0.1.12
|
305 |
+
httpcore==1.0.5
|
306 |
+
httplib2==0.21.0
|
307 |
+
httptools==0.6.1
|
308 |
+
httpx==0.27.0
|
309 |
+
huggingface-hub==0.23.2
|
310 |
+
humanfriendly==10.0
|
311 |
+
hunspell==0.5.5
|
312 |
+
hydra-slayer==0.5.0
|
313 |
+
hyperopt==0.2.7
|
314 |
+
hypertools==0.8.0
|
315 |
+
idna==3.6
|
316 |
+
igraph==0.11.5
|
317 |
+
imagecodecs==2024.6.1
|
318 |
+
imageio==2.33.1
|
319 |
+
imbalanced-learn==0.12.3
|
320 |
+
imgaug==0.4.0
|
321 |
+
importlib-metadata==6.11.0
|
322 |
+
importlib-metadata==7.0.1
|
323 |
+
importlib-resources==6.1.1
|
324 |
+
inequality==1.0.1
|
325 |
+
iniconfig==2.0.0
|
326 |
+
ipydatawidgets==4.3.5
|
327 |
+
ipykernel==6.28.0
|
328 |
+
ipyleaflet==0.19.1
|
329 |
+
ipympl==0.7.0
|
330 |
+
ipython-genutils==0.2.0
|
331 |
+
ipython-genutils==0.2.0
|
332 |
+
ipython-sql==0.5.0
|
333 |
+
ipython==8.20.0
|
334 |
+
ipyvolume==0.6.3
|
335 |
+
ipyvue==1.11.1
|
336 |
+
ipyvuetify==1.9.4
|
337 |
+
ipywebrtc==0.6.0
|
338 |
+
ipywidgets==7.7.1
|
339 |
+
isoduration==20.11.0
|
340 |
+
isort==5.13.2
|
341 |
+
isoweek==1.3.3
|
342 |
+
itsdangerous==2.2.0
|
343 |
+
jaraco.classes==3.3.0
|
344 |
+
jax-jumpy==1.0.0
|
345 |
+
jax==0.4.26
|
346 |
+
jaxlib==0.4.26.dev20240504
|
347 |
+
jedi==0.19.1
|
348 |
+
jeepney==0.8.0
|
349 |
+
jieba==0.42.1
|
350 |
+
jmespath==1.0.1
|
351 |
+
joblib==1.4.2
|
352 |
+
json5==0.9.14
|
353 |
+
jsonpatch==1.33
|
354 |
+
jsonpointer==2.4
|
355 |
+
jsonschema-specifications==2023.12.1
|
356 |
+
jsonschema==4.20.0
|
357 |
+
jupyter-console==6.6.3
|
358 |
+
jupyter-events==0.9.0
|
359 |
+
jupyter-http-over-ws==0.0.8
|
360 |
+
jupyter-leaflet==0.19.1
|
361 |
+
jupyter-lsp==1.5.1
|
362 |
+
jupyter-server-mathjax==0.2.6
|
363 |
+
jupyter-ydoc==0.2.5
|
364 |
+
jupyter_client==7.4.9
|
365 |
+
jupyter_client==8.6.0
|
366 |
+
jupyter_core==5.7.1
|
367 |
+
jupyter_server==2.12.5
|
368 |
+
jupyter_server_fileid==0.9.1
|
369 |
+
jupyter_server_proxy==4.1.0
|
370 |
+
jupyter_server_terminals==0.5.1
|
371 |
+
jupyter_server_ydoc==0.8.0
|
372 |
+
jupyterlab-lsp==5.1.0
|
373 |
+
jupyterlab-widgets==3.0.9
|
374 |
+
jupyterlab==4.2.1
|
375 |
+
jupyterlab_git==0.44.0
|
376 |
+
jupyterlab_pygments==0.3.0
|
377 |
+
jupyterlab_server==2.27.2
|
378 |
+
jupytext==1.16.0
|
379 |
+
kaggle-environments==1.14.11
|
380 |
+
kaggle==1.6.14
|
381 |
+
kagglehub==0.2.5
|
382 |
+
keras-cv==0.9.0
|
383 |
+
keras-nlp==0.12.1
|
384 |
+
keras-tuner==1.4.6
|
385 |
+
keras==3.3.3
|
386 |
+
kernels-mixer==0.0.7
|
387 |
+
keyring==24.3.0
|
388 |
+
keyrings.google-artifactregistry-auth==1.1.2
|
389 |
+
kfp-pipeline-spec==0.2.2
|
390 |
+
kfp-server-api==2.0.5
|
391 |
+
kfp==2.5.0
|
392 |
+
kiwisolver==1.4.5
|
393 |
+
kmapper==2.0.1
|
394 |
+
kmodes==0.12.2
|
395 |
+
korean-lunar-calendar==0.3.1
|
396 |
+
kornia==0.7.2
|
397 |
+
kornia_rs==0.1.3
|
398 |
+
kt-legacy==1.0.5
|
399 |
+
kubernetes==26.1.0
|
400 |
+
langcodes==3.4.0
|
401 |
+
langid==1.1.6
|
402 |
+
language_data==1.2.0
|
403 |
+
lazy_loader==0.3
|
404 |
+
learntools==0.3.4
|
405 |
+
leven==1.0.4
|
406 |
+
libclang==16.0.6
|
407 |
+
libmambapy==1.5.8
|
408 |
+
libpysal==4.9.2
|
409 |
+
librosa==0.10.2.post1
|
410 |
+
lightgbm==4.2.0
|
411 |
+
lightning-utilities==0.11.2
|
412 |
+
lime==0.2.0.1
|
413 |
+
line_profiler==4.1.3
|
414 |
+
linkify-it-py==2.0.3
|
415 |
+
llvmlite==0.41.1
|
416 |
+
llvmlite==0.42.0
|
417 |
+
lml==0.1.0
|
418 |
+
locket==1.0.0
|
419 |
+
loguru==0.7.2
|
420 |
+
lxml==5.2.2
|
421 |
+
lz4==4.3.3
|
422 |
+
mamba==1.5.8
|
423 |
+
mapclassify==2.6.1
|
424 |
+
marisa-trie==1.1.0
|
425 |
+
markdown-it-py==3.0.0
|
426 |
+
marshmallow==3.21.2
|
427 |
+
matplotlib-inline==0.1.6
|
428 |
+
matplotlib-venn==0.11.10
|
429 |
+
matplotlib==3.7.5
|
430 |
+
matplotlib==3.8.4
|
431 |
+
mccabe==0.7.0
|
432 |
+
mdit-py-plugins==0.4.0
|
433 |
+
mdurl==0.1.2
|
434 |
+
memory-profiler==0.61.0
|
435 |
+
menuinst==2.0.1
|
436 |
+
mercantile==1.2.1
|
437 |
+
mgwr==2.2.1
|
438 |
+
missingno==0.5.2
|
439 |
+
mistune==0.8.4
|
440 |
+
mizani==0.11.4
|
441 |
+
ml-dtypes==0.2.0
|
442 |
+
mlcrate==0.2.0
|
443 |
+
mlens==0.2.3
|
444 |
+
mlxtend==0.23.1
|
445 |
+
mne==1.7.0
|
446 |
+
mnist==0.2.2
|
447 |
+
momepy==0.7.0
|
448 |
+
more-itertools==10.2.0
|
449 |
+
mpld3==0.5.10
|
450 |
+
mpmath==1.3.0
|
451 |
+
msgpack==1.0.7
|
452 |
+
msgpack==1.0.8
|
453 |
+
multidict==6.0.4
|
454 |
+
multimethod==1.10
|
455 |
+
multipledispatch==1.0.0
|
456 |
+
multiprocess==0.70.16
|
457 |
+
munkres==1.1.4
|
458 |
+
murmurhash==1.0.10
|
459 |
+
mypy-extensions==1.0.0
|
460 |
+
namex==0.0.8
|
461 |
+
nb-conda-kernels==2.3.1
|
462 |
+
nb_conda==2.2.1
|
463 |
+
nbclassic==1.0.0
|
464 |
+
nbclient==0.5.13
|
465 |
+
nbclient==0.9.0
|
466 |
+
nbconvert==6.4.5
|
467 |
+
nbdime==3.2.0
|
468 |
+
nbformat==5.9.2
|
469 |
+
ndindex==1.8
|
470 |
+
nest-asyncio==1.5.8
|
471 |
+
networkx==3.2.1
|
472 |
+
nibabel==5.2.1
|
473 |
+
nilearn==0.10.4
|
474 |
+
ninja==1.11.1.1
|
475 |
+
nltk==3.2.4
|
476 |
+
nose==1.3.7
|
477 |
+
notebook==6.5.4
|
478 |
+
notebook==6.5.6
|
479 |
+
notebook_executor==0.2
|
480 |
+
notebook_shim==0.2.3
|
481 |
+
numba==0.58.1
|
482 |
+
numba==0.59.1
|
483 |
+
numexpr==2.10.0
|
484 |
+
numpy==1.26.4
|
485 |
+
nvidia-ml-py==11.495.46
|
486 |
+
nvtx==0.2.10
|
487 |
+
oauth2client==4.1.3
|
488 |
+
oauthlib==3.2.2
|
489 |
+
objsize==0.6.1
|
490 |
+
odfpy==1.4.1
|
491 |
+
olefile==0.47
|
492 |
+
onnx==1.16.1
|
493 |
+
opencensus-context==0.1.3
|
494 |
+
opencensus==0.11.4
|
495 |
+
opencv-contrib-python==4.10.0.82
|
496 |
+
opencv-python-headless==4.10.0.82
|
497 |
+
opencv-python==4.10.0.82
|
498 |
+
openpyxl==3.1.3
|
499 |
+
openslide-python==1.3.1
|
500 |
+
opentelemetry-api==1.22.0
|
501 |
+
opentelemetry-exporter-otlp-proto-common==1.22.0
|
502 |
+
opentelemetry-exporter-otlp-proto-grpc==1.22.0
|
503 |
+
opentelemetry-exporter-otlp-proto-http==1.22.0
|
504 |
+
opentelemetry-exporter-otlp==1.22.0
|
505 |
+
opentelemetry-proto==1.22.0
|
506 |
+
opentelemetry-sdk==1.22.0
|
507 |
+
opentelemetry-semantic-conventions==0.43b0
|
508 |
+
opt-einsum==3.3.0
|
509 |
+
optax==0.2.2
|
510 |
+
optimum==1.20.0
|
511 |
+
optree==0.11.0
|
512 |
+
optuna==3.6.1
|
513 |
+
orbax-checkpoint==0.5.15
|
514 |
+
ordered-set==4.1.0
|
515 |
+
orjson==3.9.10
|
516 |
+
ortools==9.4.1874
|
517 |
+
osmnx==1.9.3
|
518 |
+
overrides==7.4.0
|
519 |
+
packaging==21.3
|
520 |
+
pandas-datareader==0.10.0
|
521 |
+
pandas-profiling==3.6.6
|
522 |
+
pandas-summary==0.2.0
|
523 |
+
pandas==2.2.1
|
524 |
+
pandas==2.2.2
|
525 |
+
pandasql==0.7.3
|
526 |
+
pandocfilters==1.5.0
|
527 |
+
panel==1.4.4
|
528 |
+
papermill==2.5.0
|
529 |
+
param==2.1.0
|
530 |
+
parso==0.8.3
|
531 |
+
partd==1.4.2
|
532 |
+
path.py==12.5.0
|
533 |
+
path==16.14.0
|
534 |
+
pathos==0.3.2
|
535 |
+
pathy==0.10.3
|
536 |
+
patsy==0.5.6
|
537 |
+
pdf2image==1.17.0
|
538 |
+
peft==0.11.1
|
539 |
+
pettingzoo==1.24.0
|
540 |
+
pexpect==4.8.0
|
541 |
+
pexpect==4.9.0
|
542 |
+
phik==0.12.4
|
543 |
+
pickleshare==0.7.5
|
544 |
+
pillow==10.3.0
|
545 |
+
pip==23.3.2
|
546 |
+
pkgutil_resolve_name==1.3.10
|
547 |
+
platformdirs==3.11.0
|
548 |
+
platformdirs==4.1.0
|
549 |
+
plotly-express==0.4.1
|
550 |
+
plotly==5.18.0
|
551 |
+
plotnine==0.13.6
|
552 |
+
pluggy==1.5.0
|
553 |
+
pointpats==2.4.0
|
554 |
+
polars==0.20.31
|
555 |
+
polyglot==16.7.4
|
556 |
+
pooch==1.8.1
|
557 |
+
pox==0.3.4
|
558 |
+
ppca==0.0.4
|
559 |
+
ppft==1.7.6.8
|
560 |
+
preprocessing==0.1.13
|
561 |
+
preshed==3.0.9
|
562 |
+
prettytable==3.9.0
|
563 |
+
progressbar2==4.4.2
|
564 |
+
prometheus-client==0.19.0
|
565 |
+
promise==2.3
|
566 |
+
prompt-toolkit==3.0.42
|
567 |
+
prompt-toolkit==3.0.43
|
568 |
+
prophet==1.1.1
|
569 |
+
proto-plus==1.23.0
|
570 |
+
protobuf==3.20.3
|
571 |
+
protobuf==4.24.4
|
572 |
+
psutil==5.9.3
|
573 |
+
psutil==5.9.7
|
574 |
+
ptyprocess==0.7.0
|
575 |
+
pudb==2024.1
|
576 |
+
pure-eval==0.2.2
|
577 |
+
py-cpuinfo==9.0.0
|
578 |
+
py-spy==0.3.14
|
579 |
+
py4j==0.10.9.7
|
580 |
+
pyLDAvis==3.4.1
|
581 |
+
pyOpenSSL==23.3.0
|
582 |
+
pyaml==24.4.0
|
583 |
+
pyarrow-hotfix==0.6
|
584 |
+
pyarrow==14.0.2
|
585 |
+
pyasn1-modules==0.3.0
|
586 |
+
pyasn1==0.5.1
|
587 |
+
pybind11==2.12.0
|
588 |
+
pyclipper==1.3.0.post5
|
589 |
+
pycodestyle==2.11.1
|
590 |
+
pycosat==0.6.6
|
591 |
+
pycparser==2.21
|
592 |
+
pycryptodome==3.20.0
|
593 |
+
pyct==0.5.0
|
594 |
+
pycuda==2024.1
|
595 |
+
pydantic==2.5.3
|
596 |
+
pydantic==2.7.2
|
597 |
+
pydantic_core==2.14.6
|
598 |
+
pydantic_core==2.18.3
|
599 |
+
pydegensac==0.1.2
|
600 |
+
pydicom==2.4.4
|
601 |
+
pydocstyle==6.3.0
|
602 |
+
pydot==1.4.2
|
603 |
+
pydub==0.25.1
|
604 |
+
pyemd==1.0.0
|
605 |
+
pyerfa==2.0.1.4
|
606 |
+
pyexcel-io==0.6.6
|
607 |
+
pyexcel-ods==0.6.0
|
608 |
+
pyflakes==3.2.0
|
609 |
+
pygltflib==1.16.2
|
610 |
+
pykalman==0.9.7
|
611 |
+
pylibraft==24.4.0
|
612 |
+
pylint==3.2.2
|
613 |
+
pymc3==3.11.4
|
614 |
+
pymongo==3.13.0
|
615 |
+
pynndescent==0.5.12
|
616 |
+
pynvjitlink==0.2.3
|
617 |
+
pynvml==11.4.1
|
618 |
+
pynvrtc==9.2
|
619 |
+
pyparsing==3.1.1
|
620 |
+
pyparsing==3.1.2
|
621 |
+
pypdf==4.2.0
|
622 |
+
pyproj==3.6.1
|
623 |
+
pysal==24.1
|
624 |
+
pyshp==2.3.1
|
625 |
+
pytesseract==0.3.10
|
626 |
+
pytest==8.2.1
|
627 |
+
python-bidi==0.4.2
|
628 |
+
python-dateutil==2.9.0.post0
|
629 |
+
python-dotenv==1.0.0
|
630 |
+
python-json-logger==2.0.7
|
631 |
+
python-louvain==0.16
|
632 |
+
python-lsp-jsonrpc==1.1.2
|
633 |
+
python-lsp-server==1.11.0
|
634 |
+
python-slugify==8.0.4
|
635 |
+
python-utils==3.8.2
|
636 |
+
pythreejs==2.4.2
|
637 |
+
pytoolconfig==1.3.1
|
638 |
+
pytools==2024.1.3
|
639 |
+
pytorch-ignite==0.5.0.post2
|
640 |
+
pytorch-lightning==2.2.5
|
641 |
+
pytz==2023.3.post1
|
642 |
+
pytz==2024.1
|
643 |
+
pyu2f==0.1.5
|
644 |
+
pyviz_comms==3.0.2
|
645 |
+
pyzmq==24.0.1
|
646 |
+
pyzmq==25.1.2
|
647 |
+
qgrid==1.3.1
|
648 |
+
qtconsole==5.5.2
|
649 |
+
quantecon==0.7.2
|
650 |
+
qudida==0.0.4
|
651 |
+
raft-dask==24.4.0
|
652 |
+
rapids-dask-dependency==24.4.1a0
|
653 |
+
rasterio==1.3.10
|
654 |
+
rasterstats==0.19.0
|
655 |
+
ray-cpp==2.9.0
|
656 |
+
ray==2.9.0
|
657 |
+
referencing==0.32.1
|
658 |
+
regex==2023.12.25
|
659 |
+
requests-oauthlib==1.3.1
|
660 |
+
requests-toolbelt==0.10.1
|
661 |
+
requests==2.32.3
|
662 |
+
retrying==1.3.3
|
663 |
+
retrying==1.3.4
|
664 |
+
rfc3339-validator==0.1.4
|
665 |
+
rfc3986-validator==0.1.1
|
666 |
+
rgf-python==3.12.0
|
667 |
+
rich-click==1.8.2
|
668 |
+
rich==13.7.0
|
669 |
+
rich==13.7.1
|
670 |
+
rmm==24.4.0
|
671 |
+
rope==1.13.0
|
672 |
+
rouge==1.0.1
|
673 |
+
rpds-py==0.16.2
|
674 |
+
rsa==4.9
|
675 |
+
ruamel-yaml-conda==0.15.100
|
676 |
+
ruamel.yaml.clib==0.2.7
|
677 |
+
ruamel.yaml==0.18.5
|
678 |
+
s2sphere==0.2.5
|
679 |
+
s3fs==2024.3.1
|
680 |
+
s3transfer==0.6.2
|
681 |
+
safetensors==0.4.3
|
682 |
+
scattertext==0.1.19
|
683 |
+
scikit-image==0.22.0
|
684 |
+
scikit-learn-intelex==2024.4.0
|
685 |
+
scikit-learn==1.2.2
|
686 |
+
scikit-multilearn==0.2.0
|
687 |
+
scikit-optimize==0.10.1
|
688 |
+
scikit-plot==0.3.7
|
689 |
+
scikit-surprise==1.1.4
|
690 |
+
scipy==1.11.4
|
691 |
+
scipy==1.13.1
|
692 |
+
seaborn==0.12.2
|
693 |
+
segment_anything==1.0
|
694 |
+
segregation==2.5
|
695 |
+
semver==3.0.2
|
696 |
+
sentencepiece==0.2.0
|
697 |
+
sentry-sdk==2.3.1
|
698 |
+
setproctitle==1.3.3
|
699 |
+
setuptools-git==1.2
|
700 |
+
setuptools-scm==8.1.0
|
701 |
+
setuptools==69.0.3
|
702 |
+
shap==0.44.1
|
703 |
+
shapely==2.0.4
|
704 |
+
shellingham==1.5.4
|
705 |
+
simpervisor==1.0.0
|
706 |
+
simplejson==3.19.2
|
707 |
+
six==1.16.0
|
708 |
+
sklearn-pandas==2.2.0
|
709 |
+
slicer==0.0.7
|
710 |
+
smart-open==6.4.0
|
711 |
+
smmap==5.0.1
|
712 |
+
sniffio==1.3.0
|
713 |
+
snowballstemmer==2.2.0
|
714 |
+
snuggs==1.4.7
|
715 |
+
sortedcontainers==2.4.0
|
716 |
+
soundfile==0.12.1
|
717 |
+
soupsieve==2.5
|
718 |
+
soxr==0.3.7
|
719 |
+
spacy-legacy==3.0.12
|
720 |
+
spacy-loggers==1.0.5
|
721 |
+
spacy==3.7.3
|
722 |
+
spaghetti==1.7.5.post1
|
723 |
+
spectral==0.23.1
|
724 |
+
spglm==1.1.0
|
725 |
+
sphinx-rtd-theme==0.2.4
|
726 |
+
spint==1.0.7
|
727 |
+
splot==1.1.5.post1
|
728 |
+
spopt==0.6.0
|
729 |
+
spreg==1.4.2
|
730 |
+
spvcm==0.3.0
|
731 |
+
sqlparse==0.4.4
|
732 |
+
squarify==0.4.3
|
733 |
+
srsly==2.4.8
|
734 |
+
stable-baselines3==2.1.0
|
735 |
+
stack-data==0.6.2
|
736 |
+
stack-data==0.6.3
|
737 |
+
stanio==0.5.0
|
738 |
+
starlette==0.32.0.post1
|
739 |
+
statsmodels==0.14.1
|
740 |
+
stemming==1.0.1
|
741 |
+
stop-words==2018.7.23
|
742 |
+
stopit==1.1.2
|
743 |
+
stumpy==1.12.0
|
744 |
+
sympy==1.12.1
|
745 |
+
tables==3.9.2
|
746 |
+
tabulate==0.9.0
|
747 |
+
tangled-up-in-unicode==0.2.0
|
748 |
+
tbb==2021.12.0
|
749 |
+
tblib==3.0.0
|
750 |
+
tenacity==8.2.3
|
751 |
+
tensorboard-data-server==0.7.2
|
752 |
+
tensorboard-plugin-profile==2.15.0
|
753 |
+
tensorboard==2.15.1
|
754 |
+
tensorboardX==2.6.2.2
|
755 |
+
tensorflow-cloud==0.1.16
|
756 |
+
tensorflow-datasets==4.9.4
|
757 |
+
tensorflow-decision-forests==1.8.1
|
758 |
+
tensorflow-estimator==2.15.0
|
759 |
+
tensorflow-hub==0.16.1
|
760 |
+
tensorflow-io-gcs-filesystem==0.35.0
|
761 |
+
tensorflow-io==0.35.0
|
762 |
+
tensorflow-metadata==0.14.0
|
763 |
+
tensorflow-probability==0.23.0
|
764 |
+
tensorflow-serving-api==2.14.1
|
765 |
+
tensorflow-text==2.15.0
|
766 |
+
tensorflow-transform==0.14.0
|
767 |
+
tensorflow==2.15.0
|
768 |
+
tensorstore==0.1.60
|
769 |
+
termcolor==2.4.0
|
770 |
+
terminado==0.18.0
|
771 |
+
testpath==0.6.0
|
772 |
+
text-unidecode==1.3
|
773 |
+
textblob==0.18.0.post0
|
774 |
+
texttable==1.7.0
|
775 |
+
tf_keras==2.15.1
|
776 |
+
tfp-nightly==0.24.0.dev0
|
777 |
+
thinc==8.2.3
|
778 |
+
threadpoolctl==3.2.0
|
779 |
+
tifffile==2023.12.9
|
780 |
+
timm==1.0.3
|
781 |
+
tinycss2==1.2.1
|
782 |
+
tobler==0.11.2
|
783 |
+
tokenizers==0.19.1
|
784 |
+
toml==0.10.2
|
785 |
+
tomli==2.0.1
|
786 |
+
tomlkit==0.12.5
|
787 |
+
toolz==0.12.1
|
788 |
+
torch==2.1.2
|
789 |
+
torchaudio==2.1.2
|
790 |
+
torchdata==0.7.1
|
791 |
+
torchinfo==1.8.0
|
792 |
+
torchmetrics==1.4.0.post0
|
793 |
+
torchtext==0.16.2
|
794 |
+
torchvision==0.16.2
|
795 |
+
tornado==6.3.3
|
796 |
+
tqdm==4.66.4
|
797 |
+
traceml==1.0.8
|
798 |
+
traitlets==5.9.0
|
799 |
+
traittypes==0.2.1
|
800 |
+
transformers==4.41.2
|
801 |
+
treelite==4.1.2
|
802 |
+
truststore==0.8.0
|
803 |
+
trx-python==0.2.9
|
804 |
+
tsfresh==0.20.2
|
805 |
+
typeguard==4.1.5
|
806 |
+
typer==0.9.0
|
807 |
+
typer==0.9.4
|
808 |
+
types-python-dateutil==2.8.19.20240106
|
809 |
+
typing-inspect==0.9.0
|
810 |
+
typing-utils==0.1.0
|
811 |
+
typing_extensions==4.9.0
|
812 |
+
tzdata==2023.4
|
813 |
+
tzdata==2024.1
|
814 |
+
uc-micro-py==1.0.3
|
815 |
+
ucx-py==0.37.0
|
816 |
+
ujson==5.10.0
|
817 |
+
umap-learn==0.5.6
|
818 |
+
unicodedata2==15.1.0
|
819 |
+
update-checker==0.18.0
|
820 |
+
uri-template==1.3.0
|
821 |
+
uritemplate==3.0.1
|
822 |
+
urllib3==1.26.18
|
823 |
+
urllib3==2.1.0
|
824 |
+
urwid==2.6.12
|
825 |
+
urwid_readline==0.14
|
826 |
+
uvicorn==0.25.0
|
827 |
+
uvloop==0.19.0
|
828 |
+
vaex-astro==0.9.3
|
829 |
+
vaex-core==4.17.1
|
830 |
+
vaex-hdf5==0.14.1
|
831 |
+
vaex-jupyter==0.8.2
|
832 |
+
vaex-ml==0.18.3
|
833 |
+
vaex-server==0.9.0
|
834 |
+
vaex-viz==0.5.4
|
835 |
+
vaex==4.17.0
|
836 |
+
vec_noise==1.1.4
|
837 |
+
vecstack==0.4.0
|
838 |
+
virtualenv==20.21.0
|
839 |
+
visions==0.7.5
|
840 |
+
vowpalwabbit==9.9.0
|
841 |
+
vtk==9.3.0
|
842 |
+
wandb==0.17.0
|
843 |
+
wasabi==1.1.2
|
844 |
+
watchfiles==0.21.0
|
845 |
+
wavio==0.0.9
|
846 |
+
wcwidth==0.2.13
|
847 |
+
weasel==0.3.4
|
848 |
+
webcolors==1.13
|
849 |
+
webencodings==0.5.1
|
850 |
+
websocket-client==1.7.0
|
851 |
+
websockets==12.0
|
852 |
+
wfdb==4.1.2
|
853 |
+
whatthepatch==1.0.5
|
854 |
+
wheel==0.42.0
|
855 |
+
widgetsnbextension==3.6.6
|
856 |
+
witwidget==1.8.1
|
857 |
+
woodwork==0.31.0
|
858 |
+
wordcloud==1.9.3
|
859 |
+
wordsegment==1.3.1
|
860 |
+
wrapt==1.14.1
|
861 |
+
xarray-einstats==0.7.0
|
862 |
+
xarray==2024.5.0
|
863 |
+
xgboost==2.0.3
|
864 |
+
xvfbwrapper==0.2.9
|
865 |
+
xxhash==3.4.1
|
866 |
+
xyzservices==2024.4.0
|
867 |
+
y-py==0.6.2
|
868 |
+
yapf==0.40.2
|
869 |
+
yarl==1.9.3
|
870 |
+
yarl==1.9.4
|
871 |
+
ydata-profiling==4.6.4
|
872 |
+
yellowbrick==1.5
|
873 |
+
ypy-websocket==0.8.4
|
874 |
+
zict==3.0.0
|
875 |
+
zipp==3.17.0
|
876 |
+
zstandard==0.19.0
|
wandb/run-20240626_071628-smnm2aje/files/wandb-metadata.json
ADDED
@@ -0,0 +1,66 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"os": "Linux-5.15.133+-x86_64-with-glibc2.31",
|
3 |
+
"python": "3.10.13",
|
4 |
+
"heartbeatAt": "2024-06-26T07:16:29.539044",
|
5 |
+
"startedAt": "2024-06-26T07:16:28.965434",
|
6 |
+
"docker": null,
|
7 |
+
"cuda": null,
|
8 |
+
"args": [],
|
9 |
+
"state": "running",
|
10 |
+
"program": "kaggle.ipynb",
|
11 |
+
"codePathLocal": null,
|
12 |
+
"root": "/kaggle/working",
|
13 |
+
"host": "adc95cf38b20",
|
14 |
+
"username": "root",
|
15 |
+
"executable": "/opt/conda/bin/python3.10",
|
16 |
+
"cpu_count": 2,
|
17 |
+
"cpu_count_logical": 4,
|
18 |
+
"cpu_freq": {
|
19 |
+
"current": 2000.146,
|
20 |
+
"min": 0.0,
|
21 |
+
"max": 0.0
|
22 |
+
},
|
23 |
+
"cpu_freq_per_core": [
|
24 |
+
{
|
25 |
+
"current": 2000.146,
|
26 |
+
"min": 0.0,
|
27 |
+
"max": 0.0
|
28 |
+
},
|
29 |
+
{
|
30 |
+
"current": 2000.146,
|
31 |
+
"min": 0.0,
|
32 |
+
"max": 0.0
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"current": 2000.146,
|
36 |
+
"min": 0.0,
|
37 |
+
"max": 0.0
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"current": 2000.146,
|
41 |
+
"min": 0.0,
|
42 |
+
"max": 0.0
|
43 |
+
}
|
44 |
+
],
|
45 |
+
"disk": {
|
46 |
+
"/": {
|
47 |
+
"total": 8062.387607574463,
|
48 |
+
"used": 5693.895751953125
|
49 |
+
}
|
50 |
+
},
|
51 |
+
"gpu": "Tesla T4",
|
52 |
+
"gpu_count": 2,
|
53 |
+
"gpu_devices": [
|
54 |
+
{
|
55 |
+
"name": "Tesla T4",
|
56 |
+
"memory_total": 16106127360
|
57 |
+
},
|
58 |
+
{
|
59 |
+
"name": "Tesla T4",
|
60 |
+
"memory_total": 16106127360
|
61 |
+
}
|
62 |
+
],
|
63 |
+
"memory": {
|
64 |
+
"total": 31.357563018798828
|
65 |
+
}
|
66 |
+
}
|
wandb/run-20240626_071628-smnm2aje/files/wandb-summary.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"train/loss": 1.18, "train/grad_norm": 0.583970308303833, "train/learning_rate": 0.0, "train/epoch": 9.411764705882353, "train/global_step": 120, "_timestamp": 1719391099.5724134, "_runtime": 4910.594539403915, "_step": 21, "eval/loss": 1.2984201908111572, "eval/runtime": 127.8842, "eval/samples_per_second": 1.572, "eval/steps_per_second": 0.399, "train_runtime": 4600.4395, "train_samples_per_second": 0.437, "train_steps_per_second": 0.026, "total_flos": 775966301356032.0, "train_loss": 1.6367768367131552}
|
wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20240626_071628-smnm2aje/logs/debug.log
ADDED
@@ -0,0 +1,246 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
|
2 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
|
3 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
4 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
|
5 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
|
6 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
7 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
|
8 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
9 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
10 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
11 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
12 |
+
2024-06-26 07:16:28,967 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
|
13 |
+
2024-06-26 07:16:28,967 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
14 |
+
Traceback (most recent call last):
|
15 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
16 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
17 |
+
File "/tmp/ipykernel_34/2014566126.py", line 10, in <module>
|
18 |
+
trainer.train()
|
19 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
20 |
+
return inner_training_loop(
|
21 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
22 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
23 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
24 |
+
return self.call_event("on_train_begin", args, state, control)
|
25 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
26 |
+
result = getattr(callback, event)(
|
27 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
28 |
+
self.setup(args, state, model, **kwargs)
|
29 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
30 |
+
self._wandb.init(
|
31 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
32 |
+
wandb._sentry.reraise(e)
|
33 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
34 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
35 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
36 |
+
wi.setup(kwargs)
|
37 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
38 |
+
wandb_login._login(
|
39 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
40 |
+
wlogin.prompt_api_key()
|
41 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
42 |
+
key, status = self._prompt_api_key()
|
43 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
44 |
+
key = apikey.prompt_api_key(
|
45 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
46 |
+
key = input_callback(api_ask).strip()
|
47 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
48 |
+
value = prompt_func(prompt)
|
49 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
50 |
+
raise Abort() from None
|
51 |
+
click.exceptions.Abort
|
52 |
+
2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
53 |
+
Traceback (most recent call last):
|
54 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
55 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
56 |
+
File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
|
57 |
+
trainer.train()
|
58 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
59 |
+
return inner_training_loop(
|
60 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
61 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
62 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
63 |
+
return self.call_event("on_train_begin", args, state, control)
|
64 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
65 |
+
result = getattr(callback, event)(
|
66 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
67 |
+
self.setup(args, state, model, **kwargs)
|
68 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
69 |
+
self._wandb.init(
|
70 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
71 |
+
wandb._sentry.reraise(e)
|
72 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
73 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
74 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
75 |
+
wi.setup(kwargs)
|
76 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
77 |
+
wandb_login._login(
|
78 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
79 |
+
wlogin.prompt_api_key()
|
80 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
81 |
+
key, status = self._prompt_api_key()
|
82 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
83 |
+
key = apikey.prompt_api_key(
|
84 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
85 |
+
key = input_callback(api_ask).strip()
|
86 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
87 |
+
value = prompt_func(prompt)
|
88 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
89 |
+
raise Abort() from None
|
90 |
+
click.exceptions.Abort
|
91 |
+
2024-06-26 07:16:28,968 ERROR MainThread:34 [wandb_setup.py:_flush():78] error in wandb.init()
|
92 |
+
Traceback (most recent call last):
|
93 |
+
File "/opt/conda/lib/python3.10/site-packages/IPython/core/interactiveshell.py", line 3553, in run_code
|
94 |
+
exec(code_obj, self.user_global_ns, self.user_ns)
|
95 |
+
File "/tmp/ipykernel_34/4032920361.py", line 1, in <module>
|
96 |
+
trainer.train()
|
97 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 1885, in train
|
98 |
+
return inner_training_loop(
|
99 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer.py", line 2147, in _inner_training_loop
|
100 |
+
self.control = self.callback_handler.on_train_begin(args, self.state, self.control)
|
101 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 454, in on_train_begin
|
102 |
+
return self.call_event("on_train_begin", args, state, control)
|
103 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/trainer_callback.py", line 498, in call_event
|
104 |
+
result = getattr(callback, event)(
|
105 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 773, in on_train_begin
|
106 |
+
self.setup(args, state, model, **kwargs)
|
107 |
+
File "/opt/conda/lib/python3.10/site-packages/transformers/integrations/integration_utils.py", line 746, in setup
|
108 |
+
self._wandb.init(
|
109 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1178, in init
|
110 |
+
wandb._sentry.reraise(e)
|
111 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/analytics/sentry.py", line 155, in reraise
|
112 |
+
raise exc.with_traceback(sys.exc_info()[2])
|
113 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 1163, in init
|
114 |
+
wi.setup(kwargs)
|
115 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_init.py", line 300, in setup
|
116 |
+
wandb_login._login(
|
117 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 334, in _login
|
118 |
+
wlogin.prompt_api_key()
|
119 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 256, in prompt_api_key
|
120 |
+
key, status = self._prompt_api_key()
|
121 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 236, in _prompt_api_key
|
122 |
+
key = apikey.prompt_api_key(
|
123 |
+
File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/lib/apikey.py", line 151, in prompt_api_key
|
124 |
+
key = input_callback(api_ask).strip()
|
125 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 164, in prompt
|
126 |
+
value = prompt_func(prompt)
|
127 |
+
File "/opt/conda/lib/python3.10/site-packages/click/termui.py", line 147, in prompt_func
|
128 |
+
raise Abort() from None
|
129 |
+
click.exceptions.Abort
|
130 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug.log
|
131 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240626_071628-smnm2aje/logs/debug-internal.log
|
132 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78720aaa5b40>
|
133 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
|
134 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
|
135 |
+
config: {}
|
136 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():610] starting backend
|
137 |
+
2024-06-26 07:16:28,969 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
|
138 |
+
2024-06-26 07:16:28,971 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
139 |
+
2024-06-26 07:16:28,977 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
|
140 |
+
2024-06-26 07:16:28,989 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
|
141 |
+
2024-06-26 07:16:29,303 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
|
142 |
+
2024-06-26 07:16:29,306 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
|
143 |
+
2024-06-26 07:16:29,429 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
|
144 |
+
2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
145 |
+
|
146 |
+
2024-06-26 07:16:29,507 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
|
147 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
|
148 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
|
149 |
+
2024-06-26 07:16:45,662 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
|
150 |
+
2024-06-26 07:16:45,663 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
|
151 |
+
2024-06-26 07:16:45,672 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
|
152 |
+
2024-06-26 07:16:45,678 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-16-15_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
|
153 |
+
2024-06-26 07:18:26,851 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
154 |
+
2024-06-26 07:18:26,851 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
155 |
+
2024-06-26 07:18:50,892 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
156 |
+
2024-06-26 07:18:50,915 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
157 |
+
2024-06-26 07:18:50,915 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
158 |
+
2024-06-26 07:18:51,909 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
159 |
+
2024-06-26 07:18:51,913 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
160 |
+
2024-06-26 07:18:51,913 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
161 |
+
2024-06-26 07:18:52,853 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
162 |
+
2024-06-26 07:18:52,855 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
163 |
+
2024-06-26 07:18:52,855 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
164 |
+
2024-06-26 07:18:53,667 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
165 |
+
2024-06-26 07:18:53,670 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
166 |
+
2024-06-26 07:18:53,670 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
167 |
+
2024-06-26 07:18:54,691 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
168 |
+
2024-06-26 07:18:54,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
169 |
+
2024-06-26 07:18:54,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
170 |
+
2024-06-26 07:18:55,617 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
171 |
+
2024-06-26 07:18:55,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
172 |
+
2024-06-26 07:18:55,682 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
173 |
+
2024-06-26 07:18:56,512 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
174 |
+
2024-06-26 07:18:56,513 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
175 |
+
2024-06-26 07:18:56,513 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
176 |
+
2024-06-26 07:18:57,999 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
177 |
+
2024-06-26 07:18:58,001 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
178 |
+
2024-06-26 07:18:58,001 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
179 |
+
2024-06-26 07:18:58,997 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
180 |
+
2024-06-26 07:18:59,009 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
181 |
+
2024-06-26 07:18:59,009 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
182 |
+
2024-06-26 07:19:00,090 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
183 |
+
2024-06-26 07:19:00,092 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
184 |
+
2024-06-26 07:19:00,092 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
185 |
+
2024-06-26 07:19:01,722 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
186 |
+
2024-06-26 07:19:01,723 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
187 |
+
2024-06-26 07:19:01,723 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
188 |
+
2024-06-26 07:19:03,098 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
189 |
+
2024-06-26 07:19:03,117 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
190 |
+
2024-06-26 07:19:03,117 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
191 |
+
2024-06-26 07:19:04,117 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
192 |
+
2024-06-26 07:19:04,212 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
193 |
+
2024-06-26 07:19:04,212 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
194 |
+
2024-06-26 07:19:05,065 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
195 |
+
2024-06-26 07:19:05,097 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
196 |
+
2024-06-26 07:19:05,097 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
197 |
+
2024-06-26 07:19:06,209 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
198 |
+
2024-06-26 07:19:06,220 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
199 |
+
2024-06-26 07:19:06,220 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
200 |
+
2024-06-26 07:19:07,300 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
201 |
+
2024-06-26 07:19:07,961 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.41.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Jun26_07-19-05_adc95cf38b20', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
|
202 |
+
2024-06-26 08:35:48,386 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
203 |
+
2024-06-26 08:35:48,387 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
204 |
+
2024-06-26 08:36:11,677 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
205 |
+
2024-06-26 08:38:19,575 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
206 |
+
2024-06-26 08:38:19,575 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
207 |
+
2024-06-26 08:39:55,100 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
208 |
+
2024-06-26 08:39:55,103 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
209 |
+
2024-06-26 08:39:55,103 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
210 |
+
2024-06-26 08:40:32,017 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
211 |
+
2024-06-26 08:43:02,593 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
212 |
+
2024-06-26 08:43:02,594 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
213 |
+
2024-06-26 08:45:10,959 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
214 |
+
2024-06-26 08:45:10,963 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
215 |
+
2024-06-26 08:45:10,963 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
216 |
+
2024-06-26 08:45:13,384 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
217 |
+
2024-06-26 08:47:43,994 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
218 |
+
2024-06-26 08:47:43,994 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
219 |
+
2024-06-26 08:49:58,760 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
220 |
+
2024-06-26 08:53:33,771 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
221 |
+
2024-06-26 08:53:33,771 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
222 |
+
2024-06-26 08:54:32,055 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
223 |
+
2024-06-26 08:54:32,059 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
224 |
+
2024-06-26 08:54:32,059 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
225 |
+
2024-06-26 08:54:32,754 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
226 |
+
2024-06-26 08:58:07,756 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
227 |
+
2024-06-26 08:58:07,756 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
228 |
+
2024-06-26 09:12:46,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
229 |
+
2024-06-26 09:12:46,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
230 |
+
2024-06-26 09:12:46,669 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
231 |
+
2024-06-26 09:16:22,037 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
232 |
+
2024-06-26 09:16:22,038 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
233 |
+
2024-06-26 09:16:22,038 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
234 |
+
2024-06-26 09:16:35,228 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
235 |
+
2024-06-26 09:16:36,438 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
236 |
+
2024-06-26 09:16:36,438 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
237 |
+
2024-06-26 09:18:25,016 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
238 |
+
2024-06-26 09:18:26,213 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
239 |
+
2024-06-26 09:18:26,214 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
240 |
+
2024-06-26 09:21:21,819 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
241 |
+
2024-06-26 09:21:21,842 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
242 |
+
2024-06-26 09:21:21,842 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
243 |
+
2024-06-26 09:21:36,632 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
244 |
+
2024-06-26 09:21:36,633 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
245 |
+
2024-06-26 09:21:36,634 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
|
246 |
+
2024-06-26 09:21:37,142 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
|
wandb/run-20240626_071628-smnm2aje/run-smnm2aje.wandb
ADDED
Binary file (212 kB). View file
|
|