Back To Projects
Back To Projects

SUPA’s strategic use of graphic design experts for generative model excellence

SUPA leveraged domain-specific talent to source and label 600k stylized vector images, solving data diversity challenges

Problem statement

The client struggled to source and label 600,000 diverse vector images for training a stylistically versatile GenAI model, hampered by inconsistent vendor quality.

SOLUTION

SUPA rapidly scaled domain-expert graphic designers to curate, label, and sketch assets via iterative workflows and ML-aligned quality checks.

RESULT

SUPA delivered the dataset in three months with only 3.5% rework, enabling a 37% FID score improvement in the client’s AI model for style-diverse outputs.

Overview

SUPA’s Technical Collaboration with a Generative AI Client: Enhancing Model Robustness Through Curated Datasets

Client Profile
SUPA partnered with a U.S.-based Generative AI (GenAI) company specializing in multimodal AI systems for design asset generation. The client’s goal was to refine a generative adversarial network (GAN) architecture capable of producing vectorized design assets across industries like gaming, architecture, and e-commerce. The model’s performance hinged on a high-quality, diverse training corpus to ensure stylistic versatility and domain adaptability.

The Machine Learning Challenge: Dataset Curation for Model Generalization
Modern generative models, particularly GANs and diffusion models, require extensive, well-labeled datasets to learn latent representations of artistic styles, object geometries, and contextual semantics.

The client faced three core technical hurdles:

  1. Data Diversity & Bias Mitigation
    To avoid mode collapse—a failure case where GANs generate homogeneous outputs—the model needed exposure to a multimodal dataset. This included vector images spanning watercolor, minimalist, 3D-rendered, and hand-drawn sketch styles. Without such diversity, the model would fail to generalize across client-specified sectors.
  2. Structured Labeling for Conditional Generation
    The client’s architecture relied on conditional generation, where outputs are guided by textual or visual prompts (e.g., “watercolor icon of a castle”). Accurate semantic segmentation and labeling were critical to align image features with embeddings in the model’s latent space. Poor labeling would degrade cross-attention mechanisms in transformer-based components.
  3. Sketch-to-Asset Synthesis
    A secondary workflow required training a variational autoencoder (VAE) to convert rudimentary sketches into polished assets. This necessitated paired data: sketches (ranging from child-like doodles to professional drafts) and their corresponding vectorized outputs. The VAE’s encoder needed to learn invariant features across sketch quality levels.

Initial Vendor Limitations
Prior vendors failed to deliver datasets with sufficient entropy (diversity) and suffered from annotation inaccuracies. This introduced noise into the training pipeline, risking gradient instability during adversarial training and reducing output fidelity.

SUPA’s Machine Learning-Centric Solution
SUPA engineered a human-in-the-loop (HITL) pipeline to address the client’s technical requirements:

  1. Active Learning for Data Sourcing
    • Diversity Sampling: SUPA implemented cluster-based sampling to maximize feature space coverage. Workers sourced images from niche repositories (e.g., Behance, ArtStation) and public datasets, prioritizing underrepresented styles to minimize distributional bias.
    • Domain Expert Curation: Graphic designers on SUPA’s team applied style transfer validation, ensuring images adhered to target aesthetics (e.g., confirming watercolor textures via spectral analysis).
  2. Semantic Segmentation & Labeling
    • Hierarchical Taxonomy: A multi-label classification system was developed, tagging images by style, object class, and industry use-case. For example, a vectorized "tree" might be labeled {style: polygonal, class: flora, use-case: game design}.
    • Cross-Modal Validation: SUPA performed structured annotation of SVG vector layers, linking each component to descriptive English text. This process ensured the client’s model could interpret and manipulate individual design elements during synthesis.
  3. Sketch-Image Pair Synthesis
    • Stochastic Sketch Generation: Artists generated sketches with controlled randomness in stroke roughness and detail levels. This diversity trained the VAE to robustly encode inputs regardless of skill level.
    • Cycle Consistency Checks: Paired sketches and vector assets were validated using a pre-trained Siamese network to ensure structural similarity, filtering mismatched pairs.
  4. Pipeline Integration & Quality Assurance
    • Automated Data Cleansing: SUPA’s platform integrated Python scripts for deduplication (hashing) and outlier detection (Isolation Forest on image embeddings).
    • Versioned Dataset Exports: Data was packaged in TFRecord format with JSON metadata, compatible with the client’s PyTorch Lightning training pipeline.

Technical Outcomes & Model Impact

  • Dataset Metrics:
    • Label accuracy: 98.5% (validated via Monte Carlo sampling).
    • Sketch-to-asset pairs: 5:1 redundancy per vector image.
  • Model Performance:
    Post-training, the client reported a 37% improvement in Fréchet Inception Distance (FID) scores, indicating enhanced output realism and diversity. The model also achieved 89% accuracy in style transfer tasks (e.g., converting a 3D asset to watercolor), validated via user studies.

Why Human Expertise Matters in GenAI Training
While synthetic data generation (e.g., using Stable Diffusion) can augment datasets, domain-specific edge cases (e.g., architecturally valid geometries) require human validation. SUPA’s workforce provided three critical advantages:

  1. Bias Identification: Experts flagged underrepresented classes (e.g., Art Deco designs) for targeted sourcing.
  2. Semantic Grounding: Labelers mapped abstract client criteria (e.g., “playful style”) to concrete visual features.
  3. Iterative Feedback: Daily standups allowed rapid dataset pivots based on the client’s training loss metrics.

Conclusion
This collaboration highlights the symbiotic relationship between scalable human expertise and machine learning efficacy. By combining HITL validation with ML-driven quality checks, SUPA enabled the client to train a robust generative model capable of few-shot adaptation to novel design tasks—a critical milestone in industrial GenAI applications.

Other Projects

Discover the work we do

View All
View All

Annotation for global agritech company

SUPA scales high-quality annotation output during seasonal data surges by 170% for a global agritech company that manages over 200 million trees

problem
Aerobotics needed to handle fluctuating, large volumes of data with high-quality annotations and a 24-hour turnaround, requiring scalable and flexible workflows.
solution
SUPA’s infrastructure enabled Aerobotics to scale up to 170% in a week without building internal annotation capacity or compromising quality.
result
SUPA helped Aerobotics validate up to 200,000 tree annotations within 24 hours at 97% accuracy, ensuring year-round volume flexibility and dependable service since 2020.
200 million trees
18 countries
Agritech
Aerial Imagery
Annotation for global agritech company

Enhancing AR footwear try-on precision through high-quality segmentation

>99% accurate semantic segmentation data for ZERO10’s AR footwear try-on, enabling a precise, immersive user experience with rapid 5-week turnaround

problem
ZERO10 required enhanced precision in their footwear try-on models to deliver a more immersive and accurate AR user experience
solution
SUPA deployed its expert annotators, skilled in semantic segmentation, alongside its advanced labeling infrastructure, ensuring high-quality, scalable data annotation
result
ZERO10 successfully launched their advanced AR footwear try-on feature, powered by SUPA’s rapid delivery of segmentation datasets with >99% accuracy, enabling superior model performance and user experience
Enhancing AR footwear try-on precision through high-quality segmentation