Curl Chat Client For Multimodal#
Refer to thetrtllm-serve documentation for starting a server.
SourceNVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash 2 3# SINGLE IMAGE INFERENCE 4curlhttp://localhost:8000/v1/chat/completions\ 5-H"Content-Type: application/json"\ 6-d'{ 7 "model": "Qwen2.5-VL-3B-Instruct", 8 "messages":[{ 9 "role": "system",10 "content": "You are a helpful assistant."11 }, {12 "role": "user",13 "content": [14 {15 "type": "text",16 "text": "Describe the natural environment in the image."17 },18 {19 "type":"image_url",20 "image_url": {21 "url": "https://huggingface.co/datasets/YiYiXu/testing-images/resolve/main/seashore.png"22 }23 }24 ]25 }],26 "max_tokens": 64,27 "temperature": 028 }'2930# MULTI IMAGE INFERENCE31curlhttp://localhost:8000/v1/chat/completions\32-H"Content-Type: application/json"\33-d'{34 "model": "Qwen2.5-VL-3B-Instruct",35 "messages":[{36 "role": "system",37 "content": "You are a helpful assistant."38 }, {39 "role": "user",40 "content": [41 {42 "type": "text",43 "text":"Tell me the difference between two images"44 },45 {46 "type":"image_url",47 "image_url": {48 "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/inpaint.png"49 }50 },51 {52 "type":"image_url",53 "image_url": {54 "url": "https://huggingface.co/datasets/YiYiXu/testing-images/resolve/main/seashore.png"55 }56 }57 ]58 }],59 "max_tokens": 64,60 "temperature": 061 }'6263# SINGLE VIDEO INFERENCE64curlhttp://localhost:8000/v1/chat/completions\65-H"Content-Type: application/json"\66-d'{67 "model": "Qwen2.5-VL-3B-Instruct",68 "messages":[{69 "role": "system",70 "content": "You are a helpful assistant."71 }, {72 "role": "user",73 "content": [74 {75 "type": "text",76 "text":"Tell me what you see in the video briefly."77 },78 {79 "type":"video_url",80 "video_url": {81 "url": "https://huggingface.co/datasets/Efficient-Large-Model/VILA-inference-demos/resolve/main/OAI-sora-tokyo-walk.mp4"82 }83 }84 ]85 }],86 "max_tokens": 64,87 "temperature": 088 }'