amirabdullah19852020
commited on
Commit
•
0e68847
1
Parent(s):
1b4e8e9
Upload folder using huggingface_hub
Browse files- models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/config.json +55 -0
- models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/generation_config.json +6 -0
- models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/metrics.json +1 -0
- models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/model.safetensors +3 -0
- models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/training_args.json +1 -0
models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/config.json
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "EleutherAI/gpt-neo-125m",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPTNeoForCausalLM"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0,
|
8 |
+
"attention_layers": [
|
9 |
+
"global",
|
10 |
+
"local",
|
11 |
+
"global",
|
12 |
+
"local",
|
13 |
+
"global",
|
14 |
+
"local",
|
15 |
+
"global",
|
16 |
+
"local",
|
17 |
+
"global",
|
18 |
+
"local",
|
19 |
+
"global",
|
20 |
+
"local"
|
21 |
+
],
|
22 |
+
"attention_types": [
|
23 |
+
[
|
24 |
+
[
|
25 |
+
"global",
|
26 |
+
"local"
|
27 |
+
],
|
28 |
+
6
|
29 |
+
]
|
30 |
+
],
|
31 |
+
"bos_token_id": 50256,
|
32 |
+
"classifier_dropout": 0.1,
|
33 |
+
"embed_dropout": 0,
|
34 |
+
"eos_token_id": 50256,
|
35 |
+
"gradient_checkpointing": false,
|
36 |
+
"hidden_size": 768,
|
37 |
+
"initializer_range": 0.02,
|
38 |
+
"intermediate_size": null,
|
39 |
+
"layer_norm_epsilon": 1e-05,
|
40 |
+
"max_position_embeddings": 2048,
|
41 |
+
"model_type": "gpt_neo",
|
42 |
+
"num_heads": 12,
|
43 |
+
"num_layers": 12,
|
44 |
+
"resid_dropout": 0,
|
45 |
+
"summary_activation": null,
|
46 |
+
"summary_first_dropout": 0.1,
|
47 |
+
"summary_proj_to_labels": true,
|
48 |
+
"summary_type": "cls_index",
|
49 |
+
"summary_use_proj": true,
|
50 |
+
"torch_dtype": "bfloat16",
|
51 |
+
"transformers_version": "4.38.2",
|
52 |
+
"use_cache": true,
|
53 |
+
"vocab_size": 50257,
|
54 |
+
"window_size": 256
|
55 |
+
}
|
models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 50256,
|
4 |
+
"eos_token_id": 50256,
|
5 |
+
"transformers_version": "4.38.2"
|
6 |
+
}
|
models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_loss": 0.6949542760848999, "eval_runtime": 675.5666, "eval_samples_per_second": 12.659, "eval_steps_per_second": 1.582, "eval_rewards/chosen": 0.3397723138332367, "eval_rewards/rejected": 0.2946900725364685, "eval_rewards/accuracies": 0.5452525615692139, "eval_rewards/margins": 0.045082252472639084, "eval_logps/rejected": -157.93484497070312, "eval_logps/chosen": -164.5487823486328, "eval_logits/rejected": -16.404342651367188, "eval_logits/chosen": -16.390727996826172, "epoch": 0.25}
|
models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a3ed584a1832a548584dc0f626c8e484c8da2f25157031d134377de577d8fb4
|
3 |
+
size 250414224
|
models/hh_rlhf/gpt-neo-125m/2024-05-07_06:01/training_args.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"output_dir": "./test", "overwrite_output_dir": false, "do_train": false, "do_eval": true, "do_predict": false, "evaluation_strategy": "steps", "prediction_loss_only": false, "per_device_train_batch_size": 8, "per_device_eval_batch_size": 8, "per_gpu_train_batch_size": null, "per_gpu_eval_batch_size": null, "gradient_accumulation_steps": 1, "eval_accumulation_steps": null, "eval_delay": 0, "learning_rate": 3e-05, "weight_decay": 0.0, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "num_train_epochs": 3.0, "max_steps": 5000, "lr_scheduler_type": "linear", "lr_scheduler_kwargs": {}, "warmup_ratio": 0.0, "warmup_steps": 150, "log_level": "passive", "log_level_replica": "warning", "log_on_each_node": true, "logging_dir": "./test/runs/May07_04-27-50_a66c464a5008", "logging_strategy": "steps", "logging_first_step": true, "logging_steps": 10, "logging_nan_inf_filter": true, "save_strategy": "steps", "save_steps": 500, "save_total_limit": null, "save_safetensors": true, "save_on_each_node": false, "save_only_model": false, "no_cuda": false, "use_cpu": false, "use_mps_device": false, "seed": 42, "data_seed": null, "jit_mode_eval": false, "use_ipex": false, "bf16": true, "fp16": false, "fp16_opt_level": "O1", "half_precision_backend": "auto", "bf16_full_eval": false, "fp16_full_eval": false, "tf32": null, "local_rank": 0, "ddp_backend": null, "tpu_num_cores": null, "tpu_metrics_debug": false, "debug": [], "dataloader_drop_last": false, "eval_steps": 3000, "dataloader_num_workers": 0, "dataloader_prefetch_factor": null, "past_index": -1, "run_name": "./test", "disable_tqdm": false, "remove_unused_columns": false, "label_names": null, "load_best_model_at_end": false, "metric_for_best_model": null, "greater_is_better": null, "ignore_data_skip": false, "fsdp": [], "fsdp_min_num_params": 0, "fsdp_config": {"min_num_params": 0, "xla": false, "xla_fsdp_v2": false, "xla_fsdp_grad_ckpt": false}, "fsdp_transformer_layer_cls_to_wrap": null, "accelerator_config": {"split_batches": false, "dispatch_batches": null, "even_batches": true, "use_seedable_sampler": true}, "deepspeed": null, "label_smoothing_factor": 0.0, "optim": "adamw_hf", "optim_args": null, "adafactor": false, "group_by_length": false, "length_column_name": "length", "report_to": ["wandb"], "ddp_find_unused_parameters": null, "ddp_bucket_cap_mb": null, "ddp_broadcast_buffers": null, "dataloader_pin_memory": true, "dataloader_persistent_workers": false, "skip_memory_metrics": true, "use_legacy_prediction_loop": false, "push_to_hub": true, "resume_from_checkpoint": null, "hub_model_id": null, "hub_strategy": "every_save", "hub_token": "<HUB_TOKEN>", "hub_private_repo": false, "hub_always_push": false, "gradient_checkpointing": false, "gradient_checkpointing_kwargs": null, "include_inputs_for_metrics": false, "fp16_backend": "auto", "push_to_hub_model_id": null, "push_to_hub_organization": null, "push_to_hub_token": "<PUSH_TO_HUB_TOKEN>", "mp_parameters": "", "auto_find_batch_size": false, "full_determinism": false, "torchdynamo": null, "ray_scope": "last", "ddp_timeout": 1800, "torch_compile": false, "torch_compile_backend": null, "torch_compile_mode": null, "dispatch_batches": null, "split_batches": null, "include_tokens_per_second": false, "include_num_input_tokens_seen": false, "neftune_noise_alpha": null}
|