Nanobit commited on
Commit
cf5ae6b
β€’
1 Parent(s): 168a7a0

Feat(readme): improve docs on multi-gpu

Browse files
Files changed (1) hide show
  1. README.md +19 -2
README.md CHANGED
@@ -36,8 +36,6 @@ git clone https://github.com/OpenAccess-AI-Collective/axolotl
36
  pip3 install -e .
37
  pip3 install -U git+https://github.com/huggingface/peft.git
38
 
39
- accelerate config
40
-
41
  # finetune lora
42
  accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml
43
 
@@ -525,6 +523,21 @@ Run
525
  accelerate launch scripts/finetune.py configs/your_config.yml
526
  ```
527
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
528
  ### Inference
529
 
530
  Pass the appropriate flag to the train command:
@@ -575,6 +588,10 @@ Try set `fp16: true`
575
 
576
  Try to turn off xformers.
577
 
 
 
 
 
578
  ## Need help? πŸ™‹β™‚οΈ
579
 
580
  Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we can help you
 
36
  pip3 install -e .
37
  pip3 install -U git+https://github.com/huggingface/peft.git
38
 
 
 
39
  # finetune lora
40
  accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml
41
 
 
523
  accelerate launch scripts/finetune.py configs/your_config.yml
524
  ```
525
 
526
+ #### Multi-GPU Config
527
+
528
+ - llama FSDP
529
+ ```yaml
530
+ fsdp:
531
+ - full_shard
532
+ - auto_wrap
533
+ fsdp_config:
534
+ fsdp_offload_params: true
535
+ fsdp_state_dict_type: FULL_STATE_DICT
536
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
537
+ ```
538
+
539
+ - llama Deepspeed: append `ACCELERATE_USE_DEEPSPEED=true` in front of finetune command
540
+
541
  ### Inference
542
 
543
  Pass the appropriate flag to the train command:
 
588
 
589
  Try to turn off xformers.
590
 
591
+ > Message about accelerate config missing
592
+
593
+ It's safe to ignore it.
594
+
595
  ## Need help? πŸ™‹β™‚οΈ
596
 
597
  Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we can help you