Back To Projects
Back To Projects

SUPA’s strategic use of graphic design experts for generative model excellence

SUPA leveraged domain-specific talent to source and label 600k stylized vector images, solving data diversity challenges

Problem statement

The client struggled to source and label 600,000 diverse vector images for training a stylistically versatile GenAI model, hampered by inconsistent vendor quality.

SOLUTION

SUPA rapidly scaled domain-expert graphic designers to curate, label, and sketch assets via iterative workflows and ML-aligned quality checks.

RESULT

SUPA delivered the dataset in three months with only 3.5% rework, enabling a 37% FID score improvement in the client’s AI model for style-diverse outputs.

Overview

SUPA’s Technical Collaboration with a Generative AI Client: Enhancing Model Robustness Through Curated Datasets

Client Profile
SUPA partnered with a U.S.-based Generative AI (GenAI) company specializing in multimodal AI systems for design asset generation. The client’s goal was to refine a generative adversarial network (GAN) architecture capable of producing vectorized design assets across industries like gaming, architecture, and e-commerce. The model’s performance hinged on a high-quality, diverse training corpus to ensure stylistic versatility and domain adaptability.

The Machine Learning Challenge: Dataset Curation for Model Generalization
Modern generative models, particularly GANs and diffusion models, require extensive, well-labeled datasets to learn latent representations of artistic styles, object geometries, and contextual semantics.

The client faced three core technical hurdles:

  1. Data Diversity & Bias Mitigation
    To avoid mode collapse—a failure case where GANs generate homogeneous outputs—the model needed exposure to a multimodal dataset. This included vector images spanning watercolor, minimalist, 3D-rendered, and hand-drawn sketch styles. Without such diversity, the model would fail to generalize across client-specified sectors.
  2. Structured Labeling for Conditional Generation
    The client’s architecture relied on conditional generation, where outputs are guided by textual or visual prompts (e.g., “watercolor icon of a castle”). Accurate semantic segmentation and labeling were critical to align image features with embeddings in the model’s latent space. Poor labeling would degrade cross-attention mechanisms in transformer-based components.
  3. Sketch-to-Asset Synthesis
    A secondary workflow required training a variational autoencoder (VAE) to convert rudimentary sketches into polished assets. This necessitated paired data: sketches (ranging from child-like doodles to professional drafts) and their corresponding vectorized outputs. The VAE’s encoder needed to learn invariant features across sketch quality levels.

Initial Vendor Limitations
Prior vendors failed to deliver datasets with sufficient entropy (diversity) and suffered from annotation inaccuracies. This introduced noise into the training pipeline, risking gradient instability during adversarial training and reducing output fidelity.

SUPA’s Machine Learning-Centric Solution
SUPA engineered a human-in-the-loop (HITL) pipeline to address the client’s technical requirements:

  1. Active Learning for Data Sourcing
    • Diversity Sampling: SUPA implemented cluster-based sampling to maximize feature space coverage. Workers sourced images from niche repositories (e.g., Behance, ArtStation) and public datasets, prioritizing underrepresented styles to minimize distributional bias.
    • Domain Expert Curation: Graphic designers on SUPA’s team applied style transfer validation, ensuring images adhered to target aesthetics (e.g., confirming watercolor textures via spectral analysis).
  2. Semantic Segmentation & Labeling
    • Hierarchical Taxonomy: A multi-label classification system was developed, tagging images by style, object class, and industry use-case. For example, a vectorized "tree" might be labeled {style: polygonal, class: flora, use-case: game design}.
    • Cross-Modal Validation: SUPA performed structured annotation of SVG vector layers, linking each component to descriptive English text. This process ensured the client’s model could interpret and manipulate individual design elements during synthesis.
  3. Sketch-Image Pair Synthesis
    • Stochastic Sketch Generation: Artists generated sketches with controlled randomness in stroke roughness and detail levels. This diversity trained the VAE to robustly encode inputs regardless of skill level.
    • Cycle Consistency Checks: Paired sketches and vector assets were validated using a pre-trained Siamese network to ensure structural similarity, filtering mismatched pairs.
  4. Pipeline Integration & Quality Assurance
    • Automated Data Cleansing: SUPA’s platform integrated Python scripts for deduplication (hashing) and outlier detection (Isolation Forest on image embeddings).
    • Versioned Dataset Exports: Data was packaged in TFRecord format with JSON metadata, compatible with the client’s PyTorch Lightning training pipeline.

Technical Outcomes & Model Impact

  • Dataset Metrics:
    • Label accuracy: 98.5% (validated via Monte Carlo sampling).
    • Sketch-to-asset pairs: 5:1 redundancy per vector image.
  • Model Performance:
    Post-training, the client reported a 37% improvement in Fréchet Inception Distance (FID) scores, indicating enhanced output realism and diversity. The model also achieved 89% accuracy in style transfer tasks (e.g., converting a 3D asset to watercolor), validated via user studies.

Why Human Expertise Matters in GenAI Training
While synthetic data generation (e.g., using Stable Diffusion) can augment datasets, domain-specific edge cases (e.g., architecturally valid geometries) require human validation. SUPA’s workforce provided three critical advantages:

  1. Bias Identification: Experts flagged underrepresented classes (e.g., Art Deco designs) for targeted sourcing.
  2. Semantic Grounding: Labelers mapped abstract client criteria (e.g., “playful style”) to concrete visual features.
  3. Iterative Feedback: Daily standups allowed rapid dataset pivots based on the client’s training loss metrics.

Conclusion
This collaboration highlights the symbiotic relationship between scalable human expertise and machine learning efficacy. By combining HITL validation with ML-driven quality checks, SUPA enabled the client to train a robust generative model capable of few-shot adaptation to novel design tasks—a critical milestone in industrial GenAI applications.

Other Projects

Discover the work we do

View All
View All

Structural damage classification of civil structures for a global oil and gas conglomerate

SUPA's experts scaled client's data annotation, accurately annotating 12,000+ images to boost damage classification workflow.

problem
The Client required a specialised team to identify and assess damage of civil structures with a high accuracy of at least 90%.
solution
SUPA’s engineering experts collaborated closely with the Client by co-creating the annotation workflow and assembling a team of annotators experienced with engineering-related projects.
result
SUPA’s team of 25 annotators successfully delivered the annotations with a consistent >90% accuracy.
Structural damage classification of civil structures for a global oil and gas conglomerate

STEM Dataset

Bilingual Multimodal STEM Dataset — a curated collection of 500 Math and Physics questions in Malay and English, some enriched with relevant images.

problem
AI models often struggle with bilingual and multimodal STEM tasks due to a lack of high-quality, domain-specific datasets in languages like Malay and English.
solution
We created a curated dataset of 500 Math and Physics questions in Malay and English, complemented by a public leaderboard to benchmark AI model performance.
result
AI teams now have a reliable resource for fine-tuning and evaluating models on real-world STEM tasks, setting a new standard for bilingual and multimodal AI development.
500 high-quality Math and Physics questions
Evaluation Leaderboard
STEM-focused AI evaluation
STEM Dataset