--- license: llama2 base_model: meta-llama/Llama-2-7b-hf tags: - generated_from_trainer model-index: - name: llama2-spin-zephyr-data results: [] --- # llama2-spin-zephyr-data This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0032 - Rewards/real: -1.5322 - Rewards/generated: -20.8812 - Rewards/accuracies: 0.9992 - Rewards/margins: 19.3491 - Logps/generated: -369.1998 - Logps/real: -194.6499 - Logits/generated: 0.0527 - Logits/real: -0.5587 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-07 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - total_train_batch_size: 32 - total_eval_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rewards/real | Rewards/generated | Rewards/accuracies | Rewards/margins | Logps/generated | Logps/real | Logits/generated | Logits/real | |:-------------:|:-----:|:----:|:---------------:|:------------:|:-----------------:|:------------------:|:---------------:|:---------------:|:----------:|:----------------:|:-----------:| | 0.0098 | 0.32 | 500 | 0.0108 | -0.7647 | -14.3850 | 0.9976 | 13.6202 | -304.2368 | -186.9758 | 0.0151 | -0.6190 | | 0.0019 | 0.64 | 1000 | 0.0044 | -1.7759 | -21.6593 | 0.9984 | 19.8833 | -376.9800 | -197.0876 | 0.0480 | -0.5656 | | 0.0009 | 0.96 | 1500 | 0.0032 | -1.5322 | -20.8812 | 0.9992 | 19.3491 | -369.1998 | -194.6499 | 0.0527 | -0.5587 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2