About 140,000 results
Open links in new tab
  1. Image Layer Details - vllm/vllm-openai:v0.12.0 - Docker Hub

    vllm/vllm-openai:v0.12.0 Multi-platform Index digest sha256:6766ce0c459e24b76f3e9ba14ffc0442131ef4248c904efdcbf0d89e38be01fe

  2. vllm/docker/Dockerfile at main · vllm-project/vllm · GitHub

    # The vLLM Dockerfile is used to construct vLLM image that can be directly used # to run the OpenAI compatible server.

  3. GPU - vLLM

    There are no pre-built wheels for this device, so you must either use the pre-built Docker image or build vLLM from source. vLLM initially supports basic model inference and serving on Intel …

  4. Finding the Best Docker Image for vLLM Inference on CUDA 12.4 …

    Guides you in choosing the optimal Docker image for vLLM inference on CUDA 12.4–compatible GPUs. Compares available images and configurations to ensure you select one that …

  5. docker.io/vllm/vllm-openai:v0.12.0 - 镜像下载 | docker.io

    docker.io/vllm/vllm-openai:v0.12.0 linux/amd64 docker.io/vllm/vllm-openai:v0.12.0 - 国内下载镜像源 浏览次数:164 这是镜像描述: vllm/openai 基于 OpenAI 的 GPT-3 模型的 API 服务,支持自 …

  6. Image Layer Details - vllm/vllm-openai:v0.10.2-aarch64 - Docker

    vllm/vllm-openai:v0.10.2-aarch64 Manifest digest sha256:447307d0fcdf2756cc3cbf5ad7c2bc3172d9d6a8c07de5865195402164215191 OS/ARCH

  7. Releases · vllm-project/vllm - GitHub

    Breaking Changes: This release includes PyTorch 2.9.0 upgrade (CUDA 12.9), V0 deprecations including xformers backend, and scheduled removals - please review the changelog carefully.

  8. Installation — vLLM

    If either you have a different CUDA version or you want to use an existing PyTorch installation, you need to build vLLM from source. See below for instructions.

  9. Installation and Setup | vllm-project/vllm | DeepWiki

    Nov 30, 2025 · vLLM can be installed through three primary methods: pre-built wheels via pip, Docker containers, or from source compilation. The build system uses CMake for compiling …

  10. GitHub - vllm-project/vllm: A high-throughput and memory …

    We explored vLLM and LMCache inference and low-resource language adaptation with speakers from Embedded LLM, AMD, and Red Hat. Please find the meetup slides here.

  11. GPU - vLLM

    There are no pre-built wheels for this device, so you must either use the pre-built Docker image or build vLLM from source. vLLM initially supports basic model inference and serving on Intel …

  12. [Bug]: cuda version of `vllm/vllm-openai:latest` older than k8s …

    Jul 30, 2025 · 🐛 Describe the bug I am running into No platform detected / CUDA initialization error. Error 803: system has unsupported display driver / cuda driver combination. Is the vllm/vllm …