Curl Chat Client For Multimodal#

Refer to thetrtllm-serve documentation for starting a server.

SourceNVIDIA/TensorRT-LLM.

 1#! /usr/bin/env bash 2 3# SINGLE IMAGE INFERENCE 4curlhttp://localhost:8000/v1/chat/completions\ 5-H"Content-Type: application/json"\ 6-d'{ 7        "model": "Qwen2.5-VL-3B-Instruct", 8        "messages":[{ 9            "role": "system",10            "content": "You are a helpful assistant."11        }, {12            "role": "user",13            "content": [14                {15                    "type": "text",16                    "text": "Describe the natural environment in the image."17                },18                {19                    "type":"image_url",20                    "image_url": {21                        "url": "https://huggingface.co/datasets/YiYiXu/testing-images/resolve/main/seashore.png"22                    }23                }24            ]25        }],26        "max_tokens": 64,27        "temperature": 028    }'2930# MULTI IMAGE INFERENCE31curlhttp://localhost:8000/v1/chat/completions\32-H"Content-Type: application/json"\33-d'{34        "model": "Qwen2.5-VL-3B-Instruct",35        "messages":[{36            "role": "system",37            "content": "You are a helpful assistant."38        }, {39            "role": "user",40            "content": [41                {42                    "type": "text",43                    "text":"Tell me the difference between two images"44                },45                {46                    "type":"image_url",47                    "image_url": {48                        "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/inpaint.png"49                    }50                },51                {52                    "type":"image_url",53                    "image_url": {54                        "url": "https://huggingface.co/datasets/YiYiXu/testing-images/resolve/main/seashore.png"55                    }56                }57            ]58        }],59        "max_tokens": 64,60        "temperature": 061    }'6263# SINGLE VIDEO INFERENCE64curlhttp://localhost:8000/v1/chat/completions\65-H"Content-Type: application/json"\66-d'{67        "model": "Qwen2.5-VL-3B-Instruct",68        "messages":[{69            "role": "system",70            "content": "You are a helpful assistant."71        }, {72            "role": "user",73            "content": [74                {75                    "type": "text",76                    "text":"Tell me what you see in the video briefly."77                },78                {79                    "type":"video_url",80                    "video_url": {81                        "url": "https://huggingface.co/datasets/Efficient-Large-Model/VILA-inference-demos/resolve/main/OAI-sora-tokyo-walk.mp4"82                    }83                }84            ]85        }],86        "max_tokens": 64,87        "temperature": 088    }'