Skip to content

Commit

Permalink
[Docs] Mention PubMedGPT
Browse files Browse the repository at this point in the history
  • Loading branch information
tridao committed Dec 16, 2022
1 parent 04c4c61 commit ece8f05
Showing 1 changed file with 5 additions and 0 deletions.
5 changes: 5 additions & 0 deletions usage.md
Original file line number Diff line number Diff line change
Expand Up @@ -45,6 +45,11 @@ yields the fastest BERT training on cloud instances in MLPerf training 2.0 (June

## Language model training & inference

- [PubMedGPT 2.7B](https://crfm.stanford.edu/2022/12/15/pubmedgpt.html), a
domain-specific LLM for biomedicine, by Stanford CRFM, trained on
[MosaicML](https://www.mosaicml.com/blog/introducing-pubmed-gpt) Cloud. Just
using FlashAttention nearly halves the total training time.

- Meta's
[AITemplate](https://ai.facebook.com/blog/gpu-inference-engine-nvidia-amd-open-source/)
uses FlashAttention as part of their approach to speed up Transformer
Expand Down

0 comments on commit ece8f05

Please sign in to comment.