flash_llama / README.md
Birchlabs's picture
Update README.md
ce7b47b
---
license: apache-2.0
---
This repository houses a fork of [`togethercomputer/LLaMA-2-7B-32K`](https://huggingface.co/togethercomputer/LLaMA-2-7B-32K)'s [`modeling_flash_llama.py`](https://huggingface.co/togethercomputer/LLaMA-2-7B-32K/blob/main/modeling_flash_llama.py), with a [fix for padding of attention weights](https://huggingface.co/togethercomputer/LLaMA-2-7B-32K/discussions/17) merged into it.