Skip to content

Commit 0d0583a

Browse files
authored
Update README.md (#2668)
1 parent be17881 commit 0d0583a

File tree

1 file changed

+13
-7
lines changed

1 file changed

+13
-7
lines changed

README.md

+13-7
Original file line numberDiff line numberDiff line change
@@ -17,13 +17,19 @@ TensorRT-LLM
1717
<div align="left">
1818

1919
## Latest News
20-
* [2024/12/10] ⚡ Llama 3.3 70B from AI at Meta is accelerated by TensorRT-LLM. 🌟 State-of-the-art model on par with Llama 3.1 405B for reasoning, math, instruction following and tool use. Explore the preview
21-
[➡️ link](https://build.nvidia.com/meta/llama-3_3-70b-instruct)
20+
* [2025/01/07] 🌟 Getting Started with TensorRT-LLM
21+
[➡️ link](https://www.youtube.com/watch?v=TwWqPnuNHV8)
22+
23+
* [2025/01/04] ⚡Boost Llama 3.3 70B Inference Throughput 3x with NVIDIA TensorRT-LLM Speculative Decoding
24+
[➡️ link](https://developer.nvidia.com/blog/boost-llama-3-3-70b-inference-throughput-3x-with-nvidia-tensorrt-llm-speculative-decoding/)
2225
<div align="center">
23-
<img src="https://media.licdn.com/dms/image/v2/D4E10AQEMmE0xCJ6QvQ/image-shrink_800/image-shrink_800/0/1733853602424?e=1734912000&v=beta&t=mjW-P7skGnxGSgMx07IWczyYYC_05RYypB2mEmbgqZQ" width="50%">
26+
<img src="https://developer-blogs.nvidia.com/wp-content/uploads/2024/12/three-llamas-wearing-goggles.png" width="50%">
2427
<div align="left">
2528

26-
* [2024/12/03] 🌟 Boost your AI hashtag#inference throughput by up to 3.6x. We now support speculative decoding and tripling token throughput with our NVIDIA TensorRT-LLM. Perfect for your generative AI apps. ⚡Learn how in this technical deep dive
29+
* [2024/12/10] ⚡ Llama 3.3 70B from AI at Meta is accelerated by TensorRT-LLM. 🌟 State-of-the-art model on par with Llama 3.1 405B for reasoning, math, instruction following and tool use. Explore the preview
30+
[➡️ link](https://build.nvidia.com/meta/llama-3_3-70b-instruct)
31+
32+
* [2024/12/03] 🌟 Boost your AI inference throughput by up to 3.6x. We now support speculative decoding and tripling token throughput with our NVIDIA TensorRT-LLM. Perfect for your generative AI apps. ⚡Learn how in this technical deep dive
2733
[➡️ link](https://nvda.ws/3ZCZTzD)
2834

2935
* [2024/12/02] Working on deploying ONNX models for performance-critical applications? Try our NVIDIA Nsight Deep Learning Designer ⚡ A user-friendly GUI and tight integration with NVIDIA TensorRT that offers:
@@ -52,6 +58,9 @@ TensorRT-LLM
5258
🙌 Enter for a chance to win prizes including an NVIDIA® GeForce RTX™ 4080 SUPER GPU, DLI credits, and more🙌
5359
[➡️ link](https://developer.nvidia.com/llamaindex-developer-contest)
5460

61+
<details close>
62+
<summary>Previous News</summary>
63+
5564
* [2024/10/28] 🏎️🏎️🏎️ NVIDIA GH200 Superchip Accelerates Inference by 2x in Multiturn Interactions with Llama Models
5665
[➡️ link](https://developer.nvidia.com/blog/nvidia-gh200-superchip-accelerates-inference-by-2x-in-multiturn-interactions-with-llama-models/)
5766

@@ -65,9 +74,6 @@ TensorRT-LLM
6574
* [2024/10/07] 🚀🚀🚀Optimizing Microsoft Bing Visual Search with NVIDIA Accelerated Libraries
6675
[➡️ link](https://developer.nvidia.com/blog/optimizing-microsoft-bing-visual-search-with-nvidia-accelerated-libraries/)
6776

68-
<details close>
69-
<summary>Previous News</summary>
70-
7177
* [2024/09/29] 🌟 AI at Meta PyTorch + TensorRT v2.4 🌟 ⚡TensorRT 10.1 ⚡PyTorch 2.4 ⚡CUDA 12.4 ⚡Python 3.12
7278
[➡️ link](https://github.com/pytorch/TensorRT/releases/tag/v2.4.0)
7379

0 commit comments

Comments
 (0)