Skip to content
@Inferact

Inferact

Our mission is to grow vLLM as the world's AI inference engine and accelerate AI progress by making inference cheaper and faster.

Popular repositories Loading

  1. vllm-large-scale-serving vllm-large-scale-serving Public

    Forked from vllm-project/vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

    Python

Repositories

Showing 1 of 1 repositories

Top languages

Python

Most used topics

Loading…