Skip to content

Conversation

@lee101
Copy link
Contributor

@lee101 lee101 commented Jun 4, 2025

Summary

  • add vLLM based inference helper with min-probability and sentence stopping
  • integrate vLLM into the FastAPI server when available
  • make tests/conftest resilient if httpx is missing
  • add basic unit test for vLLM inference

Testing

  • pytest -q (fails: ModuleNotFoundError: No module named 'cachetools')

https://chatgpt.com/codex/tasks/task_e_683fe5b06bf083338fb1ba3540b415dc

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants