Because somebody on Reddit asked how could he caption a dataset for Qwen Image and mentain consistancy I made a small ComfyUI workflow that uses Qwen 2.5 VL 7B Instruct to autocaption the images in a folder, name them, caption them and save them all in another folder. It should be straightforward to use but you will have to manage the missing nodes and models yourself
Hello, I used your workflow and it worked great, thank you. However, I encountered a problem when adding tags to my second training set. My first training set was 1024×1024, and its images were not compressed or cropped during this process.
But my second training set has resolutions like 768×1024, 1024×768, and 832×1226, and all the images were forcibly cropped to 1024×1024. I would like to know the cause of this problem and how to solve it.
Hello, I used your workflow and it worked great, thank you. However, I encountered a problem when adding tags to my second training set. My first training set was 1024×1024, and its images were not compressed or cropped during this process.
But my second training set has resolutions like 768×1024, 1024×768, and 832×1226, and all the images were forcibly cropped to 1024×1024. I would like to know the cause of this problem and how to solve it.