from vllm import LLM from vllm.assets.image import ImageAsset def run_llava(): llm = LLM(model="llava-hf/llava-1.5-7b-hf") prompt = "USER: \nWhat is the content of this image?\nASSISTANT:" image = ImageAsset("stop_sign").pil_image outputs = llm.generate({ "prompt": prompt, "multi_modal_data": { "image": image }, }) for o in outputs: generated_text = o.outputs[0].text print(generated_text) if __name__ == "__main__": run_llava()