| --- |
| title: Multipack (Sample Packing) |
| description: Multipack is a technique to pack multiple sequences into a single batch to increase training throughput. |
| --- |
|
|
| ## Visualization of Multipack with Flash Attention |
|
|
| Because Flash Attention simply drops the attention mask, we do not need to |
| construct a 4d attention mask. We only need to concatenate the sequences into |
| a single batch and let flash attention know where each new sequence begins. |
|
|
|
|
| 4k context, bsz =4, |
| each character represents 256 tokens |
| X represents a padding token |
|
|
| ``` |
| 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 |
| [[ A A A A A A A A A A A ] |
| B B B B B B ] |
| C C C C C C C ] |
| D D D D ]] |
|
|
| [[ E E E E E E E E ] |
| [ F F F F ] |
| [ G G G ] |
| [ H H H H ]] |
|
|
| [[ I I I ] |
| [ J J J ] |
| [ K K K K K] |
| [ L L L ]] |
| ``` |
|
|
| after padding to longest input in each step |
| ``` |
| 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 |
| [[ A A A A A A A A A A A ] |
| B B B B B B X X X X X X ] |
| C C C C C C C X X X X ] |
| D D D D X X X X X X X ]] |
|
|
| [[ E E E E E E E E ] |
| [ F F F F X X X X ] |
| [ G G G X X X X X ] |
| [ H H H H X X X X ]] |
|
|
| [[ I I I X X ] |
| [ J J J X X ] |
| [ K K K K K ] |
| [ L L L X X ]] |
| ``` |
|
|
| w packing ( note it's the same effective number of tokens per step, but a true bsz of 1) |
| ``` |
| 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 |
| [[ A A A A A A A A A A A B B B B B |
| B C C C C C C C D D D D E E E E |
| E E E E F F F F F G G G H H H H |
| I I I J J J J K K K K K L L L X ]] |
| ``` |
|
|
| cu_seqlens: |
| [[ 0, 11, 17, 24, 28, 36, 41 44, 48, 51, 55, 60, 64]] |
|
|
|
|
| ## Multipack without Flash Attention |
|
|
| Multipack can still be achieved without Flash attention, but with lower packing |
| efficiency as we are not able to join multiple batches into a single batch due to |
| context length limits without flash attention. We can use either Pytorch's Scaled |
| Dot Product Attention implementation or native Pytorch attention implementation |
| along with [4d attention masks](https://github.com/huggingface/transformers/pull/27539) |
| to pack sequences together and avoid cross attention. |
|
|
| <img src="./images/4d-mask.png" alt="axolotl" width="800"> |
|
|