random command line tools for deep learning
pip install dlcalc
or
git clone https://github.com/jfc4050/dlcalc
cd dlcalc
pip install -e .
After this you should have access to the command line tools described below. Some
people may need to add --user
to their pip install command for them to properly
go under $PATH
.
Calculator for estimating various performance characteristics of 3D parallel transformer model training:
- memory consumption
- pipeline bubble
- communication overhead
- compute intensity
- etc..
This calculator is focused primarily on pretraining, so you won't find calculations for things like LoRA or RLHF/DPO/etc.
For more details run:
3dtrn -h
We've include a sample config you can use to see what the output looks like:
3dtrn examples/llama3_70b.yaml
We recommend pairing this with a profiler of your choice (NVIDIA Nsight Systems and PyTorch Profiler Traces are two good ones), and checking that what you see in the profiler is in the same ballpark as the theoretical values estimated by the calculator. If they are way off, you now know where to spend your investigation/debugging time.
NOTE: If you look at the sample config, you'll see that it takes an instance type which is used to derive various hardware specifications like intra/inter node bandwidth, theoretical FLOPS per device, number of accelerators per node, etc. You can check hardware.py to see what instance types are supported, you'll have to add it there if the instance type you're interested in isn't represented.
Given a kubernetes pod name prefix for some compute cluster, retrieve AWS network topology info and plot. For more details run:
topoviz -h
Evaluates how optimal a given training job's physical topology is (in terms of network hops in each DP ring). For more details run:
topoeval -h
Topology-aware instance selection and rank assignment for maximally efficient DP communication. For more details run:
topoassign -h
Pretty self explanatory, for more details run:
sps2tpd -h
If you're not familiar with what Model Flops Utilization (MFU) means, refer to Google's PaLM paper. Otherwise pretty self explanatory, for more details run:
sps2mfu -h
Gives a human-readable summarization of keys, values, and tensor shapes in a given (PyTorch) model checkpoint. For more details run:
ckpt-summarize -h
install development dependencies
pip install -e .[dev]
static checks can be run with
bash checks