Loading request...
User wants an easy setup for vLLM that automatically installs and runs SOTA models (like llamacpp/ikllama/kobold/vllm) with optimal GPU and RAM settings, and a wrapper for Claude code subscription through a CC proxy.
@Teknium @outsource_ havent had time to isntall and try hermes yet, though before opening it i would wish there was an easy setup, 1. install + run llamacpp/ikllama/kobold/vllm sota model with ur gpu + ram limit on auto good settings. 2. wrapper for claude code subscription through CC proxy(legal?)