Update README.md
Browse files
README.md
CHANGED
|
@@ -53,7 +53,7 @@ You can change aggregate strategies using the `--agg_strategy` flag, such as `sa
|
|
| 53 |
|
| 54 |
## Train Step 1 ##
|
| 55 |
```
|
| 56 |
-
accelerate launch --config_file=./accelerate_configs/deepspeed_zero2.yaml run_wsi.py --learning_rate 1e-4 --
|
| 57 |
--gpu 2 --train_batch_size 4 --eval_batch_size 2 --max_seq_length 512 \
|
| 58 |
--agg_strategy gmm,longnet --embed_dim 512 --vision_adaptor False --hierachical_token True --hierachical_adaptor True\
|
| 59 |
--n_heads 32,16,8 --llm_requires_grad False --resume_from_checkpoint False \
|
|
@@ -67,7 +67,7 @@ accelerate launch --config_file=./accelerate_configs/deepspeed_zero2.yaml run_ws
|
|
| 67 |
|
| 68 |
## Train Step 2 ##
|
| 69 |
```
|
| 70 |
-
accelerate launch --config_file=./accelerate_configs/deepspeed_zero2.yaml run_wsi.py --
|
| 71 |
--gpu 2 --train_batch_size 8 --eval_batch_size 2 --max_seq_length 256 \
|
| 72 |
--agg_strategy gmm,longnet --embed_dim 512 --vision_adaptor False --hierachical_token True --hierachical_adaptor True\
|
| 73 |
--n_heads 32,16,8 --llm_requires_grad True --resume_from_checkpoint False \
|
|
|
|
| 53 |
|
| 54 |
## Train Step 1 ##
|
| 55 |
```
|
| 56 |
+
accelerate launch --config_file=./accelerate_configs/deepspeed_zero2.yaml run_wsi.py --learning_rate 1e-4 --num_train_epochs 20 --warmup_steps 1000\
|
| 57 |
--gpu 2 --train_batch_size 4 --eval_batch_size 2 --max_seq_length 512 \
|
| 58 |
--agg_strategy gmm,longnet --embed_dim 512 --vision_adaptor False --hierachical_token True --hierachical_adaptor True\
|
| 59 |
--n_heads 32,16,8 --llm_requires_grad False --resume_from_checkpoint False \
|
|
|
|
| 67 |
|
| 68 |
## Train Step 2 ##
|
| 69 |
```
|
| 70 |
+
accelerate launch --config_file=./accelerate_configs/deepspeed_zero2.yaml run_wsi.py --num_train_epochs 5 --warmup_steps 1000\
|
| 71 |
--gpu 2 --train_batch_size 8 --eval_batch_size 2 --max_seq_length 256 \
|
| 72 |
--agg_strategy gmm,longnet --embed_dim 512 --vision_adaptor False --hierachical_token True --hierachical_adaptor True\
|
| 73 |
--n_heads 32,16,8 --llm_requires_grad True --resume_from_checkpoint False \
|