Creative Testing at Scale: How Top Indonesian Brands Optimize Ad Performance - banner image
Back to Insights
Performance MarketingFebruary 18, 202610 min read

Creative Testing at Scale: How Top Indonesian Brands Optimize Ad Performance

By Maju Mapan Digital

In 2025, creative quality is the single largest driver of paid advertising performance. As platform algorithms become increasingly sophisticated at finding the right audience, the variable that separates top-performing campaigns from mediocre ones is the creative itself. Meta's own research shows that creative quality accounts for 56% of the variance in ad auction outcomes — more than audience targeting, bidding strategy, or placement combined. Yet most Indonesian brands still treat creative as a production exercise ("make it look nice") rather than a performance science ("test, measure, iterate").

Why Creative Testing Matters More Than Ever

Three trends have elevated creative testing from "nice to have" to "essential":

  • Algorithm dependency: As Meta's Advantage+ and Google's Performance Max shift toward automated targeting, your creative is the primary signal the algorithm uses to find the right audience. Better creative = better algorithmic matching = lower CPA.
  • Audience fatigue: Indonesian social media users are exposed to 6,000-10,000 ad impressions daily. Creative fatigue sets in faster than ever — a winning ad today can become a losing ad within 2-3 weeks.
  • Rising CPMs: As more advertisers compete for Indonesian audiences, CPMs are rising 15-20% year-over-year. Creative efficiency is the primary lever for maintaining performance against rising costs.

The Creative Testing Framework

Level 1: Concept Testing

Concept testing evaluates fundamentally different creative approaches — different messaging angles, value propositions, or emotional appeals. This is the highest-impact level of testing because concept changes can swing performance by 50-200%.

For an Indonesian skincare brand, concept tests might include: "Clinical results" (data-driven, ingredient-focused), "Real transformations" (before/after, user testimonials), "Insider secrets" (expert tips, Korean beauty trends), "Affordable luxury" (premium positioning at accessible prices).

Run concept tests with equal budget across 3-5 concepts, measuring CPA and ROAS after reaching statistical significance (typically 1,000+ impressions per concept). Kill underperformers, scale winners.

Level 2: Format Testing

Once you identify winning concepts, test different creative formats to find the optimal presentation:

  • Static image vs. video vs. carousel: Performance varies dramatically by category. For Indonesian beauty brands, video outperforms static by 2-3x. For Indonesian B2B, static images often outperform video.
  • UGC-style vs. polished production: Test creator-filmed content against studio-produced content. Our data shows UGC-style outperforms for TikTok by 65% but polished wins on Instagram by 20%.
  • Short (6-15s) vs. medium (15-30s) vs. long (30-60s): Optimal length varies by platform and funnel stage. Top-of-funnel awareness ads perform best at 6-15 seconds, while retargeting ads can be 30-60 seconds because the audience already has context.

Level 3: Element Testing

The most granular level — testing individual elements within a proven concept and format:

  • Hook (first 3 seconds): The single most impactful element to test. A stronger hook can improve thumbstop rate (the percentage of people who stop scrolling) by 2-5x. Our creative team always produces 3-5 hook variations for every winning concept.
  • CTA (Call to Action): "Shop Now" vs. "Learn More" vs. "Get Yours" vs. "Limited Time" can swing CTR by 20-40%.
  • Copy overlay: Testing different text overlays, benefit callouts, and price positioning within the same visual framework.
  • Audio/music: On TikTok and Reels, trending audio vs. custom audio vs. voiceover significantly impacts engagement.

The Production Machine

Systematic creative testing requires a production pipeline that can output 15-25 new creative assets per week per platform. This sounds overwhelming, but it's achievable with the right system:

  • Modular creative system: Design templates where hooks, body content, and CTAs can be swapped independently. This turns 5 hooks × 3 bodies × 3 CTAs into 45 variations from 11 components.
  • Creator network: Maintain relationships with 10-20 micro-creators who can produce UGC-style content on brief. Each creator produces 3-5 raw takes per brief, which can be edited into multiple variations.
  • AI-assisted production: Use AI tools for scriptwriting, thumbnail generation, and copy variation — then refine with human creative direction.
  • Centralized creative library: Tag and organize every creative by concept, format, element, and performance data. This historical database becomes increasingly valuable for identifying patterns and predicting winners.

Reading the Data

Common creative testing mistakes in Indonesian campaigns:

  • Not waiting for significance: Making decisions based on 200 impressions is noise, not signal. Wait for 1,000+ impressions and stable metrics before declaring winners.
  • Optimizing for the wrong metric: High CTR doesn't always mean high ROAS. An entertaining ad might get clicks from curiosity-seekers who don't convert. Always tie creative testing back to business outcomes (CPA, ROAS, CLV).
  • Ignoring creative fatigue: Every winning creative has a lifespan. Monitor frequency (number of times a user sees the ad) and refresh creative when CPA starts rising — typically at frequency 3-5 for Indonesian audiences.
  • Testing too many variables simultaneously: Changing hook, format, CTA, and audience in the same test makes it impossible to attribute performance differences. Isolate one variable per test.

Frequently Asked Questions

How much should I spend on creative testing vs. scaling winners?

Allocate 20-30% of your paid media budget to creative testing and 70-80% to scaling proven winners. The testing budget funds your innovation pipeline — without it, your "winners" become stale, and performance degrades over time. For a brand spending Rp 100 million monthly on paid social, this means Rp 20-30 million dedicated to testing new concepts, formats, and elements.

How do I know when a creative is "winning"?

A creative is winning when it consistently meets or exceeds your target KPI (CPA or ROAS) after reaching statistical significance. In practical terms: if your target CPA is Rp 50,000, a creative that achieves Rp 35,000 CPA over 2,000+ impressions is a clear winner. Scale it to 70-80% of budget. A creative at Rp 48,000 is marginal — keep testing variations. Above Rp 65,000 is a clear loser — kill it.

Should creative testing happen in a separate campaign or within existing campaigns?

Best practice is to use a dedicated testing campaign with controlled settings (specific audience, consistent budget, manual placements for clean data). Winners from the testing campaign are then graduated into your main scaling campaigns. This prevents testing from disrupting the performance of your proven campaigns. Meta's campaign budget optimization (CBO) with ad-level cost cap can also enable testing within scaling campaigns, but it's harder to read results cleanly.

3