Skip to content

Batch inference support #5

@quetzalsly

Description

@quetzalsly

RTX PRO 6000
59gb vram usage per inference call
90 seconds per output image

python3 '/root/instance/JoyAI-Image/inference.py' --ckpt-root '/root/instance/models/JoyAI_Image_Edit' --prompt 'Move the camera.
- Camera rotation: Yaw 180°, Pitch 30°.
- Camera zoom: unchanged.
- Keep the 3D scene static; only change the viewpoint.' --image '/root/instance/in/kik_test.png' --output '/root/instance/out/kik/a.png' --seed '8954' --steps '50' --guidance-scale '4' --neg-prompt '' --basesize '1024' --hsdp-shard-dim '1'
- ```

Currently the model loads and unloads to vram on every call.
I would like to ask if its possible to implement batch mode.

--prompts_file '/root/instance/in/prompts.json'

Maybe some other tricks to speed it up?

Thanks.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions