This repository has been archived by the owner on Feb 7, 2025. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 7
/
Copy pathindex.html
156 lines (142 loc) · 5.82 KB
/
index.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
<!DOCTYPE html>
<html lang="en"><head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge">
<meta name="viewport" content="width=device-width, initial-scale=1"><!-- Begin Jekyll SEO tag v2.8.0 -->
<title>vLLM Blog</title>
<meta name="generator" content="Jekyll v4.3.3" />
<meta property="og:title" content="vLLM Blog" />
<meta name="author" content="© 2025. vLLM Team. All rights reserved." />
<meta property="og:locale" content="en_US" />
<link rel="canonical" href="https://blog.vllm.ai/" />
<meta property="og:url" content="https://blog.vllm.ai/" />
<meta property="og:site_name" content="vLLM Blog" />
<meta property="og:type" content="website" />
<meta name="twitter:card" content="summary" />
<meta property="twitter:title" content="vLLM Blog" />
<script type="application/ld+json">
{"@context":"https://schema.org","@type":"WebSite","author":{"@type":"Person","name":"© 2025. vLLM Team. All rights reserved."},"headline":"vLLM Blog","name":"vLLM Blog","url":"https://blog.vllm.ai/"}</script>
<!-- End Jekyll SEO tag -->
<link rel="stylesheet" href="/assets/css/style.css"><link type="application/atom+xml" rel="alternate" href="https://blog.vllm.ai/feed.xml" title="vLLM Blog" /><script async src="https://www.googletagmanager.com/gtag/js?id=G-9C5R3JR3QS"></script>
<script>
window['ga-disable-G-9C5R3JR3QS'] = window.doNotTrack === "1" || navigator.doNotTrack === "1" || navigator.doNotTrack === "yes" || navigator.msDoNotTrack === "1";
window.dataLayer = window.dataLayer || [];
function gtag(){window.dataLayer.push(arguments);}
gtag('js', new Date());
gtag('config', 'G-9C5R3JR3QS');
</script>
</head>
<body><header class="site-header">
<div class="wrapper"><a class="site-title" rel="author" href="/">vLLM Blog</a></div>
</header>
<main class="page-content" aria-label="Content">
<div class="wrapper">
<div class="home">
<ul class="post-list"><li>
<span class="post-meta">Jan 27, 2025</span>
<h3>
<a class="post-link" href="/2025/01/27/v1-alpha-release.html">
vLLM V1: A Major Upgrade to vLLM's Core Architecture
</a>
</h3></li><li>
<span class="post-meta">Jan 27, 2025</span>
<h3>
<a class="post-link" href="/2025/01/27/intro-to-llama-stack-with-vllm.html">
Introducing vLLM Inference Provider in Llama Stack
</a>
</h3></li><li>
<span class="post-meta">Jan 21, 2025</span>
<h3>
<a class="post-link" href="/2025/01/21/stack-release.html">
High Performance and Easy Deployment of vLLM in K8S with “vLLM production-stack”
</a>
</h3></li><li>
<span class="post-meta">Jan 14, 2025</span>
<h3>
<a class="post-link" href="/2025/01/14/struct-decode-intro.html">
Structured Decoding in vLLM: a gentle introduction
</a>
</h3></li><li>
<span class="post-meta">Jan 10, 2025</span>
<h3>
<a class="post-link" href="/2025/01/10/vllm-2024-wrapped-2025-vision.html">
vLLM 2024 Retrospective and 2025 Vision
</a>
</h3></li><li>
<span class="post-meta">Jan 10, 2025</span>
<h3>
<a class="post-link" href="/2025/01/10/dev-experience.html">
Installing and Developing vLLM with Ease
</a>
</h3></li><li>
<span class="post-meta">Oct 23, 2024</span>
<h3>
<a class="post-link" href="/2024/10/23/vllm-serving-amd.html">
Serving LLMs on AMD MI300X: Best Practices
</a>
</h3></li><li>
<span class="post-meta">Oct 17, 2024</span>
<h3>
<a class="post-link" href="/2024/10/17/spec-decode.html">
How Speculative Decoding Boosts vLLM Performance by up to 2.8x
</a>
</h3></li><li>
<span class="post-meta">Sep 5, 2024</span>
<h3>
<a class="post-link" href="/2024/09/05/perf-update.html">
vLLM v0.6.0: 2.7x Throughput Improvement and 5x Latency Reduction
</a>
</h3></li><li>
<span class="post-meta">Jul 25, 2024</span>
<h3>
<a class="post-link" href="/2024/07/25/lfai-perf.html">
vLLM’s Open Governance and Performance Roadmap
</a>
</h3></li><li>
<span class="post-meta">Jul 23, 2024</span>
<h3>
<a class="post-link" href="/2024/07/23/llama31.html">
Announcing Llama 3.1 Support in vLLM
</a>
</h3></li><li>
<span class="post-meta">Nov 14, 2023</span>
<h3>
<a class="post-link" href="/2023/11/14/notes-vllm-vs-deepspeed.html">
Notes on vLLM v.s. DeepSpeed-FastGen
</a>
</h3></li><li>
<span class="post-meta">Jun 20, 2023</span>
<h3>
<a class="post-link" href="/2023/06/20/vllm.html">
vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention
</a>
</h3></li></ul>
</div>
</div>
</main><footer class="site-footer h-card">
<data class="u-url" href="/"></data>
<div class="wrapper">
<div class="footer-col-wrapper">
<div class="footer-col">
<!-- <p class="feed-subscribe">
<a href="https://blog.vllm.ai/feed.xml">
<svg class="svg-icon orange">
<use xlink:href="/assets/minima-social-icons.svg#rss"></use>
</svg><span>Subscribe</span>
</a>
</p> -->
<ul class="contact-list">
<li class="p-name">© 2025. vLLM Team. All rights reserved.</li>
<li><a href="https://github.com/vllm-project/vllm">https://github.com/vllm-project/vllm</a></li>
</ul>
</div>
<div class="footer-col">
<p></p>
</div>
</div>
<div class="social-links"><ul class="social-media-list"></ul>
</div>
</div>
</footer>
</body>
</html>