Skip to content

Commit

Permalink
πŸ”₯πŸ”₯πŸ”₯[DeepSeek-V3] DeepSeek-V3 Technical Report(@deepseek-ai) (#109)
Browse files Browse the repository at this point in the history
  • Loading branch information
DefTruth authored Jan 3, 2025
1 parent b5c9834 commit 12416b5
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -69,6 +69,7 @@ Awesome-LLM-Inference: A curated list of [πŸ“™Awesome LLM Inference Papers with
|2024.07|πŸ”₯πŸ”₯[**FlashAttention-3**] FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision(@TriDao etc) |[[pdf]](https://tridao.me/publications/flash3/flash3.pdf)|[[flash-attention]](https://github.com/Dao-AILab/flash-attention) ![](https://img.shields.io/github/stars/Dao-AILab/flash-attention.svg?style=social)|⭐️⭐️ |
|2024.07|πŸ”₯πŸ”₯[**MInference 1.0**] MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via Dynamic Sparse Attention(@Microsoft) |[[pdf]](https://arxiv.org/pdf/2407.02490)|[[MInference 1.0]](https://github.com/microsoft/MInference) ![](https://img.shields.io/github/stars/microsoft/MInference.svg?style=social)|⭐️⭐️ |
|2024.11|πŸ”₯πŸ”₯πŸ”₯[**Star-Attention: 11x~ speedup**] Star Attention: Efficient LLM Inference over Long Sequences(@NVIDIA)|[[pdf]](https://arxiv.org/pdf/2411.17116)|[[Star-Attention]](https://github.com/NVIDIA/Star-Attention) ![](https://img.shields.io/github/stars/NVIDIA/Star-Attention.svg?style=social)|⭐️⭐️ |
|2024.12|πŸ”₯πŸ”₯πŸ”₯[**DeepSeek-V3**] DeepSeek-V3 Technical Report(@deepseek-ai) | [[pdf]](https://github.com/deepseek-ai/DeepSeek-V3/blob/main/DeepSeek_V3.pdf) | [[DeepSeek-V3]](https://github.com/deepseek-ai/DeepSeek-V3) ![](https://img.shields.io/github/stars/deepseek-ai/DeepSeek-V3.svg?style=social) | ⭐️⭐️ |

### πŸ“–DP/MP/PP/TP/SP/CP Parallelism ([©️backπŸ‘†πŸ»](#paperlist))
<div id="DP-MP-PP-TP-SP-CP"></div>
Expand Down

0 comments on commit 12416b5

Please sign in to comment.