Skip to content

Commit dfaa2ba

Browse files
committed
docs: fix the header of the scaling test table (deepmodeling#4507)
Fix deepmodeling#4494. <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit - **Documentation** - Updated the parallel training documentation for TensorFlow and PyTorch to enhance clarity. - Expanded explanations on parallel training processes and data loading utilities. - Introduced a flowchart to illustrate data flow and modified the scaling tests table format for better understanding. <!-- end of auto-generated comment: release notes by coderabbit.ai --> Signed-off-by: Jinzhe Zeng <jinzhe.zeng@rutgers.edu> (cherry picked from commit bd2395c)
1 parent 35d5577 commit dfaa2ba

File tree

1 file changed

+8
-7
lines changed

1 file changed

+8
-7
lines changed

doc/train/parallel-training.md

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -27,13 +27,14 @@ In some cases, it won't work well when scaling the learning rate by worker count
2727
### Scaling test
2828

2929
Testing `examples/water/se_e2_a` on an 8-GPU host, linear acceleration can be observed with the increasing number of cards.
30-
31-
| Num of GPU cards | Seconds every 100 samples | Samples per second | Speed up |
32-
| ---------------- | ------------------------- | ------------------ | -------- |
33-
| 1 | 1.4515 | 68.89 | 1.00 |
34-
| 2 | 1.5962 | 62.65\*2 | 1.82 |
35-
| 4 | 1.7635 | 56.71\*4 | 3.29 |
36-
| 8 | 1.7267 | 57.91\*8 | 6.72 |
30+
In this example, the number of samples per batch on a single GPU card ({ref}`batch_size <training/training_data/batch_size>`) is set to `1`.
31+
32+
| Num of GPU cards | Samples per batch | Seconds every 100 batches | Samples per second | Speed up |
33+
| ---------------- | ----------------- | ------------------------- | ------------------ | -------- |
34+
| 1 | 1 | 1.4515 | 68.89 | 1.00 |
35+
| 2 | 2 | 1.5962 | 62.65\*2 | 1.82 |
36+
| 4 | 4 | 1.7635 | 56.71\*4 | 3.29 |
37+
| 8 | 8 | 1.7267 | 57.91\*8 | 6.72 |
3738

3839
### How to use
3940

0 commit comments

Comments
 (0)