Loading request...
The user requests LoRA support for the Async VLLM worker in verl.
currently, LoRA seems to be supported only for the vLLMRollout (sync. one) https://github.com/volcengine/verl/blob/7737bf06e5c598a8ff4c05d07b713450d4cf308d/verl/workers/sharding_manager/fsdp_vllm.py#L266 inference_engine for the async worker doesn't have the attribute llm_engine, and thus an error is raised. is anyone working on this issue? if not, I'd love to contribute!