A Python test suite for evaluating video description capabilities using the Qwen3-VL-32B-Thinking model.
This project tests the Qwen VL (Vision-Language) model's ability to precisely describe and analyze video content. The test script encodes local video files and sends them to the Qwen API for detailed analysis.
- Video encoding to base64 for API compatibility
- Streaming response output
- Detailed video content analysis
- Product demonstration detection and timestamping
- Visual quality and content potential scoring
- Python 3.8+
- DashScope API Key (from Alibaba Cloud Model Studio)
- Clone the repository:
git clone https://github.com/ZeiProX76/cotignacMafia.git
cd cotignacMafia- Create and activate a virtual environment:
python3 -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate- Install dependencies:
pip install -r requirements.txt- Set up your environment variables:
cp .env.example .env
# Edit .env and add your DASHSCOPE_API_KEYRun the video description test:
python test_video_description.pyThe script will:
- Load your video file
- Encode it to base64
- Send it to the Qwen3-VL model
- Stream and display the detailed analysis
Edit the video_path variable in test_video_description.py to point to your video file:
video_path = "/path/to/your/video.mp4"This project uses the Qwen API in OpenAI-compatible mode:
- Base URL:
https://dashscope-intl.aliyuncs.com/compatible-mode/v1 - Model:
qwen3-vl-32b-thinking
MIT
ZeiProX76