Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Would it be possible to run something like vLLM or TensortRT-llm with tinfoil?


We’re already using vllm as our inference server for our standard models. We can run whatever inference server for custom deployments.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: