Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
27 commits
Select commit Hold shift + click to select a range
4eedff1
chore: Do not package any binaries in the wheel
grahamking Jul 22, 2025
90d8a89
Missing the project.scripts section
grahamking Jul 22, 2025
3612ac7
Remove hatch_build.py from the containers
grahamking Jul 22, 2025
09be098
Also add optional dependencies
grahamking Jul 22, 2025
08eb11e
Merge branch 'main' into gk-no-bins
nv-anants Jul 23, 2025
21fc2c4
update dockerfiles
nv-anants Jul 23, 2025
32b21e1
fixes and formatting
nv-anants Jul 23, 2025
9169db2
remove gds path for nixl trtllm
nv-anants Jul 23, 2025
b07b6bc
fix vllm install
nv-anants Jul 23, 2025
a93c2bc
more vllm fixes
nv-anants Jul 23, 2025
d6c4626
add back sdk for mypy
nv-anants Jul 23, 2025
80c0788
Merge branch 'main' into gk-no-bins
nv-anants Jul 23, 2025
2f3f162
fix pre merge
nv-anants Jul 23, 2025
2f3e3fd
fix pre commit
nv-anants Jul 23, 2025
aeba695
remove req files
nv-anants Jul 23, 2025
a989a8f
remove sdk
nv-anants Jul 23, 2025
c32cdb6
add sdk to dockerignore
nv-anants Jul 23, 2025
ec0babb
Revert "add sdk to dockerignore"
nv-anants Jul 23, 2025
7351aed
Revert "remove sdk"
nv-anants Jul 23, 2025
57f44bf
remove gds path for other dockerfiles
nv-anants Jul 23, 2025
68e3c75
remove [nixl] option
nv-anants Jul 24, 2025
0a02846
Merge branch 'main' into gk-no-bins
nv-anants Jul 24, 2025
6aea572
Add TensorRT-LLM to option dependency and corresponding instructions
tanmayv25 Jul 25, 2025
bd1459c
Merge branch 'main' into tanmayv-trtllm-install
nv-anants Jul 25, 2025
7e8d033
Update pyproject.toml
nv-anants Jul 25, 2025
c8a875a
Add a todo to update the docker file to ai-dynamo[trtllm]
tanmayv25 Jul 25, 2025
e84930e
Merge branch 'main' into tanmayv-trtllm-install
tanmayv25 Jul 25, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
34 changes: 33 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -79,6 +79,7 @@ uv pip install pip
# Choose one
uv pip install "ai-dynamo[sglang]"
uv pip install "ai-dynamo[vllm]"
uv pip install "ai-dynamo[trtllm]"
uv pip install "ai-dynamo[llama_cpp]" # CPU, see later for GPU
```

Expand Down Expand Up @@ -173,7 +174,38 @@ You can pass any sglang flags directly to this worker, see https://docs.sglang.a

# TRT-LLM

This currently requires a container TODO ADD THE DOCS PLZ THANK YOU
It is recommended to use [NGC PyTorch Container](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch) for running TensorRT-LLM engine.

> [!Note]
> Ensure that you select a PyTorch container image version that matches the version of TensorRT-LLM you are using.
> For example, if you are using `tensorrt-llm==1.0.0rc4`, use the PyTorch container image version `25.05`.
> To find the correct PyTorch container version for your desired `tensorrt-llm` release, visit the [TensorRT-LLM Dockerfile.multi](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docker/Dockerfile.multi) on GitHub. Switch to the branch that matches your `tensorrt-llm` version, and look for the `BASE_TAG` line to identify the recommended PyTorch container tag.

> [!Important]
> Launch container with the following additional settings `--shm-size=1g --ulimit memlock=-1`

## Install prerequites
```
# Optional step: Only required for Blackwell and Grace Hopper
pip3 install torch==2.7.1 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128

sudo apt-get -y install libopenmpi-dev
```

> [!Tip]
> You can learn more about these prequisites and known issues with TensorRT-LLM pip based installation [here](https://nvidia.github.io/TensorRT-LLM/installation/linux.html).

## Install dynamo
```
uv pip install --upgrade pip setuptools && uv pip install ai-dynamo[trtllm]
```

Run the backend/worker like this:
```
python -m dynamo.trtllm --help
```

To specify which GPUs to use set environment variable `CUDA_VISIBLE_DEVICES`.

# llama.cpp

Expand Down
2 changes: 2 additions & 0 deletions container/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -94,6 +94,8 @@ TRTLLM_USE_NIXL_KVCACHE_EXPERIMENTAL="0"

# TensorRT-LLM PyPI index URL
TENSORRTLLM_INDEX_URL="https://pypi.python.org/simple"
# TODO: Remove the version specification from here and use the ai-dynamo[trtllm] package.
# Need to update the Dockerfile.tensorrt_llm to use the ai-dynamo[trtllm] package.
DEFAULT_TENSORRTLLM_PIP_WHEEL="tensorrt-llm==1.0.0rc0"
TENSORRTLLM_PIP_WHEEL=""

Expand Down
5 changes: 5 additions & 0 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -59,6 +59,11 @@ keywords = ["llm", "genai", "inference", "nvidia", "distributed", "dynamo"]
Repository = "https://github.com/ai-dynamo/dynamo.git"

[project.optional-dependencies]
trtllm =[
"uvloop",
"tensorrt-llm==1.0.0rc4"
]

vllm = [
"uvloop",
"nixl",
Expand Down
Loading