CVPR 2026 · HiGen Workshop

CatalogStitch

Dimension-Aware and Occlusion-Preserving Object Compositing for Catalog Image Generation

Sanyam Jain· Pragya Kandari· Manit Singhal· He Zhang· Soo Ye Kim
Adobe
Paper Dataset

Abstract

Generative object compositing methods have shown remarkable ability to seamlessly insert objects into scenes. However, when applied to real-world catalog image generation, these methods require tedious manual intervention: users must carefully adjust masks when product dimensions differ, and painstakingly restore occluded elements post-generation. We present CatalogStitch, a set of model-agnostic techniques that automate these corrections, enabling user-friendly content creation. Our dimension-aware mask computation algorithm automatically adapts the target region to accommodate products with different dimensions; users simply provide a product image and background, without manual mask adjustments. Our occlusion-aware hybrid restoration method guarantees pixel-perfect preservation of occluding elements, eliminating post-editing workflows. We additionally introduce CatalogStitch-Eval, a 58-example benchmark covering aspect-ratio mismatch and occlusion-heavy catalog scenarios, together with supplementary PDF and HTML viewers. We evaluate our techniques with three state-of-the-art compositing models (ObjectStitch, OmniPaint, and InsertAnything), demonstrating consistent improvements across diverse catalog scenarios. By reducing manual intervention and automating tedious corrections, our approach transforms generative compositing into a practical, human-friendly tool for production catalog workflows.

Dimension-Aware Mask Compositing

Products retain their native proportions with our adapted mask, while freeform and bounding-box masks cause stretching and distortion. Results consistent across all three compositing models.

a
Shoe replacement with different proportions
Background
Product
Freeform
BBox
Ours
InsertAnything
BG Product Freeform BBox Ours
OmniPaint
BG Product Freeform BBox Ours
ObjectStitch
BG Product Freeform BBox Ours
b
Helmet replacement with aspect ratio mismatch
InsertAnything
BG Product Freeform BBox Ours
OmniPaint
BG Product Freeform BBox Ours
ObjectStitch
BG Product Freeform BBox Ours
c
Product swap in lifestyle setting
InsertAnything
BG Product Freeform BBox Ours
OmniPaint
BG Product Freeform BBox Ours
ObjectStitch
BG Product Freeform BBox Ours

Occlusion-Aware Restoration

Foreground occluders are perfectly preserved after compositing via exact pixel restoration. Comparing before and after our restoration step across mask types and compositing models.

a
Sofa with foreground furniture
Background
Product
Occluders
Freeform
BBox
Ours
InsertAnything — Before
BG Product Occluders Before FF Before BB Before Ours
InsertAnything — After Restoration
BG Product Occluders After FF After BB After Ours
ObjectStitch — Before
BG Product Occluders Before FF Before BB Before Ours
ObjectStitch — After Restoration
BG Product Occluders After FF After BB After Ours
b
Scene with decorative elements
InsertAnything — Before
BG Product Occluders Before FF Before BB Before Ours
InsertAnything — After Restoration
BG Product Occluders After FF After BB After Ours
ObjectStitch — Before
BG Product Occluders Before FF Before BB Before Ours
ObjectStitch — After Restoration
BG Product Occluders After FF After BB After Ours
c
Generated scene with occluding objects
InsertAnything — Before
BG Product Occluders Before FF Before BB Before Ours
InsertAnything — After Restoration
BG Product Occluders After FF After BB After Ours
ObjectStitch — Before
BG Product Occluders Before FF Before BB Before Ours
ObjectStitch — After Restoration
BG Product Occluders After FF After BB After Ours

Quantitative Results

Evaluated on CatalogStitch-Eval across five metrics. Our techniques consistently improve every baseline model.

Method AR Error ↓ Occ. PSNR ↑ FID ↓ CLIP ↑ DINO ↑
OmniPaint31.07137.1583.0368.50
OmniPaint + Ours4.57135.7986.1172.99
ObjectStitch30.9711.60101.5590.2785.76
ObjectStitch + Ours5.0526.8491.5290.6288.09
InsertAnything29.9813.33105.9990.2382.63
InsertAnything + Ours3.9227.5477.7292.6888.30

AR Error drops from ~30% to ~4–5% uniformly across all models. InsertAnything + Ours achieves best results on all five metrics.

CatalogStitch-Eval Benchmark

A challenging evaluation benchmark specifically designed for catalog image compositing, covering both core failure modes.

35
Dimension Mismatch Scenes

Products with significantly different aspect ratios — tall lamps replacing wide tables, square bags replacing rectangular clutches, etc.

23
Occlusion Scenes

Products partially occluded by 1–2 foreground elements — plants, vases, side tables, lamps, and decorative objects.

Explore the Benchmark

Dataset Overview Metadata, images, and masks Dimension-Aware Results 35 aspect-ratio mismatch examples Occlusion Results 23 occlusion-heavy examples

Citation

@inproceedings{catalogstitch, title = {CatalogStitch: Dimension-Aware and Occlusion-Preserving Object Compositing for Catalog Image Generation}, author = {Sanyam Jain and Pragya Kandari and Manit Singhal and He Zhang and Soo Ye Kim}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, workshop = {HiGen: Human-Interactive Generation and Editing}, year = {2026} }