Skip to content

Commit 2b30e12

Browse files
committed
add logo
Signed-off-by: kerthcet <[email protected]>
1 parent 69a8728 commit 2b30e12

File tree

3 files changed

+12
-3
lines changed

3 files changed

+12
-3
lines changed

README.md

Lines changed: 12 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,13 @@
1-
# llmaz
1+
<p align="center">
2+
<picture>
3+
<source media="(prefers-color-scheme: dark)" srcset="./docs/assets/logo.png">
4+
<img alt="llmaz" src="./docs/assets/logo.png" width=55%>
5+
</picture>
6+
</p>
7+
8+
<h3 align="center">
9+
Easy, advanced inference platform for large language models on Kubernetes
10+
</h3>
211

312
[![stability-alpha](https://img.shields.io/badge/stability-alpha-f4d03f.svg)](https://github.com/mkenney/software-guides/blob/master/STABILITY-BADGES.md#alpha)
413
[![GoReport Widget]][GoReport Status]
@@ -17,8 +26,8 @@
1726

1827
## Feature Overview
1928

20-
- **User Friendly**: People can quick deploy a LLM service with minimal configurations.
21-
- **High Performance**: llmaz supports a wide range of advanced inference backends for high performance, like [vLLM](https://github.com/vllm-project/vllm), [SGLang](https://github.com/sgl-project/sglang), [llama.cpp](https://github.com/ggerganov/llama.cpp). Find the full list of supported backends [here](./docs/support-backends.md).
29+
- **Easy of Use**: People can quick deploy a LLM service with minimal configurations.
30+
- **Broad Backend Support**: llmaz supports a wide range of advanced inference backends for high performance, like [vLLM](https://github.com/vllm-project/vllm), [SGLang](https://github.com/sgl-project/sglang), [llama.cpp](https://github.com/ggerganov/llama.cpp). Find the full list of supported backends [here](./docs/support-backends.md).
2231
- **Scaling Efficiency (WIP)**: llmaz works smoothly with autoscaling components like [Cluster-Autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) or [Karpenter](https://github.com/kubernetes-sigs/karpenter) to support elastic scenarios.
2332
- **Accelerator Fungibility (WIP)**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance.
2433
- **SOTA Inference (WIP)**: llmaz supports the latest cutting-edge researches like [Speculative Decoding](https://arxiv.org/abs/2211.17192) or [Splitwise](https://arxiv.org/abs/2311.18677) to run on Kubernetes.

docs/assets/.DS_Store

0 Bytes
Binary file not shown.

docs/assets/logo.png

98.9 KB
Loading

0 commit comments

Comments
 (0)