TensorRT-LLMs/tests/unittest/llmapi/apps/_test_openai_misc.py
xiweny 6979afa6f2
test: reorganize tests folder hierarchy (#2996)
1. move TRT path tests to 'trt' folder
2. optimize some import usage
2025-03-27 12:07:53 +08:00

53 lines
1.3 KiB
Python

import openai
import pytest
import requests
from tensorrt_llm.version import __version__ as VERSION
from ..test_llm import get_model_path
from .openai_server import RemoteOpenAIServer
@pytest.fixture(scope="module")
def model_name():
return "llama-models-v2/TinyLlama-1.1B-Chat-v1.0"
@pytest.fixture(scope="module", params=[None, 'pytorch'])
def backend(request):
return request.param
@pytest.fixture(scope="module")
def server(model_name: str, backend: str):
model_path = get_model_path(model_name)
args = ["--max_beam_width", "4"]
if backend is not None:
args.append("--backend")
args.append(backend)
with RemoteOpenAIServer(model_path, args) as remote_server:
yield remote_server
@pytest.fixture(scope="module")
def client(server: RemoteOpenAIServer):
return server.get_client()
def test_version(server: RemoteOpenAIServer):
version_url = server.url_for("version")
response = requests.get(version_url)
assert response.status_code == 200
assert response.json()["version"] == VERSION
def test_health(server: RemoteOpenAIServer):
health_url = server.url_for("health")
response = requests.get(health_url)
assert response.status_code == 200
def test_model(client: openai.OpenAI, model_name: str):
model = client.models.list().data[0]
assert model.id == model_name.split('/')[-1]