trt-llm-as-openai-windows
janhq/trt-llm-as-openai-windows
This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows instead of cloud.
0stars
Forks
1
Open issues
0
Watchers
0
Size
1.1 MB
PythonOther
Created: Mar 7, 2024
Updated: Jul 23, 2024
Last push: Mar 8, 2024
ArchivedFork