Skip to content

HuaweiCloudDeveloper/vllm-image

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 

Repository files navigation

vLLM large model service

English | 简体中文

Table of Contents

Repository Introduction

vLLM is a fast and easy-to-use library for LLM inference and serving. Originally developed by the Sky Computing Lab at UC Berkeley, vLLM has evolved into a community-driven project jointly maintained by academia and industry.

The open-source image product vLLM large model service provided by this project comes pre-installed with the vLLM inference framework and its related runtime environment, along with deployment templates. Follow the usage guide to easily enjoy an efficient "out-of-the-box" experience.

System Requirements:

  • CPU: 8 vCPUs or higher
  • RAM: 16GB or larger
  • Disk: At least 40GB

Prerequisites

Register a Huawei Account and Activate Huawei Cloud

Image Description

Image Specifications Feature Description Notes
vLLM-v0.8.3-Kunpeng Deployed on Kunpeng Cloud Server + Ubuntu 24.04 64bit / Huawei Cloud EulerOS 2.0 64bit

Getting Help

How to Contribute

  • Fork this repository and submit pull requests
  • Synchronize updates to README.md based on your open-source image information

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •