1. Overview: What the Integration Is & the Value It Delivers
Prism brings together Fospha’s daily, full-funnel measurement with Smartly.io’s Predictive Budget Allocation (PBA) — automating budget shifts across channels, campaigns, and ad sets to drive incremental impact every day.
It’s the only daily MMM integrated directly with Smartly’s PBA — combining scientific measurement accuracy with operational automation.
Why it Matters
Modern marketing teams face three consistent challenges:
- Too many decisions, not enough hours — high-value optimisation opportunities get missed. 
- Siloed or biased data — platform-reported metrics exaggerate performance in their own ecosystem. 
- Unstable signals — frequent resets in platform learning create inconsistent outcomes. 
Prism solves these by turning Fospha’s unbiased, model-based data into Smartly’s optimisation engine — so spend shifts automatically toward the campaigns that actually drive new customers and incremental revenue.
In short:
- More growth, less lift — PBA executes the “fifth task on your list” every day. 
- Smarter budgets — spend moves toward campaigns proven to drive true business results. 
- Daily confidence — MMM-level accuracy refreshed every day for stable, cross-channel guidance 
2. How It Works: The Closed Feedback Loop
Prism operates as a 4-step continuous loop between Fospha and Smartly:
- Fospha provides the data 
 Fospha shares 14 days of attributed performance data — conversions, new-customer conversions, and revenue — mapped to campaign and ad IDs through automated matching.
 → Source: daily S2S feed directly from Fospha.
- Smartly ingests the data 
 A secure server-to-server (S2S) API transfers Fospha data into Smartly’s dashboard as a new optimisation signal for PBA.
- Smartly reallocates budget 
 Smartly’s Predictive Budget Allocation automatically redistributes spend across campaigns and channels based on Fospha’s MMM-led performance data — optimising toward your chosen KPI (ROAS, revenue, conversions, or new-customer ROAS).
- Fospha validates the results 
 Fospha measures the incremental effect of those budget changes and reports Test vs Control outcomes, creating a closed feedback loop that strengthens over time.
3. Best Practice: Identifying the Right Test Opportunity + Guardrails
To ensure Prism delivers consistent, measurable value, start with well-defined pilot tests that balance volume, stability, and learning potential.
How to Spot a Strong Test Opportunity
- Start with conversion-focused campaigns. - Immediate results are most visible in conversion-driven activity (e.g. evergreen prospecting or DPA campaigns). 
- Avoid mixing funnel stages (Awareness, Consideration, Conversion) in one pool; PBA works best when all campaigns share a single optimisation objective. 
 
- Ensure sufficient volume and headroom. - Choose campaigns with enough conversions to generate a stable learning signal (≈5–10× expected daily CPA budget in the pool). 
- Use Fospha’s Incremental Forecasting outputs to identify campaigns with untapped growth potential. 
 
- Exclude your very best performers. - Don’t risk early downside on top performers; start with campaigns that are stable but not mission-critical. 
 
- Structure pools around markets or clusters. - For country tests: keep each market isolated (e.g. Germany as a standalone pool) or use clusters of comparable markets (e.g. Nordics, Benelux) to learn allocation across regions with similar dynamics. 
- Avoid mixing markets with radically different baseline CVR/ROAS, as PBA will over-weight cheaper ones. 
 
- Explore cross-platform where relevant. - Single-market, cross-platform pools (e.g. Germany across Meta, TikTok, Pinterest) are high-value tests for identifying the optimal channel mix. 
- Start with one platform if you want a controlled pilot, then expand to cross-platform once you’ve validated results. 
 
- Choose the right KPI. - Decide upfront which Fospha KPI to optimise toward: Conversions, Revenue, ROAS, or New-Customer ROAS. 
- One KPI per pool keeps optimisation clear; at most two can be combined (e.g. ROAS + Revenue for value-based outcomes). 
 
- Align with org structure and ownership. - Pools should reflect budget ownership in the team (e.g. by channel, market, or region). 
- Assign a clear budget owner for each pool to manage alignment with business goals. 
 
Examples of Strong Test Candidates
- Always-on conversion campaigns that today only get monthly adjustments but could benefit from continuous optimisation. 
- Cross-platform spend in a single high-value market (e.g. Germany across Meta/TikTok/Pinterest) to test channel allocation. 
- Clusters of mid-performing regions (e.g. Nordics, Benelux) where manual optimisation is minimal and PBA can surface hidden opportunities. 
- Underperforming or overlooked regions where teams lack time to actively optimise, but Fospha signals can guide reallocation. 
Guardrails to configure
| Parameter | Recommended Setting | Why It Matters | 
| Budget Mode | Fixed Daily | Safest first step; stable total spend | 
| Min/Max Caps | 10% min / 60% max | Prevents starvation or over-concentration | 
| Exploration Budget | 10% | Allows testing new allocations safely | 
| Optimisation Cadence | Meta/TikTok: every 2 daysPinterest/Snap: every 3–4 days | Avoid manual edits between cycles | 
| Separate Pools | Prospecting / Retargeting / App Installs | Keeps funnel stages independent | 
| KPI per Pool | ROAS, Revenue, Conversions, or New-Customer ROAS | Ensures clear optimisation goal | 
Avoid these pitfalls:
- Editing budgets manually during the first 7 days (resets learning) 
- Mixing prospecting & retargeting in one pool 
- No caps or inconsistent objectives 
- Changing KPI mid-test — start a new 4-week test instead 
4. Designing the Pilot Test
4.1 Picking the Right Campaign to Test
- Start with conversion-focused, always-on campaigns. 
- Ensure campaigns share a common KPI (conversion, revenue, or ROAS). 
- For country tests: split ad set = country. 
- For channel tests: group campaigns across Meta/TikTok/Pinterest. 
- Keep creative comparable; limit to 2–3 creative themes so PBA is not confounded by asset mix. 
4.2 Budget Pools Setup Best Practice
| Setting | Recommendation | Why | 
| Pool budget size | ≥ 5–10× expected daily CPA | Ensures enough signals for 7-day learning phase | 
| Min / Max allocation | Min ≥ 10%, Max ≤ 60% per channel/campaign | Prevents starvation or over-concentration | 
| Exploration budget | Start at 10% | Allows PBA to test new allocations without risk | 
| Campaign structure | Disable Meta CBO when pooling ad sets | Gives PBA budget control at pooled level | 
| Objective buckets | Separate Prospecting, Retargeting, App Installs | Prevents PBA over-optimising into remarketing | 
4.3 Timeline
| Phase | Duration | Key actions | 
| Warm-up | Days 1–7 | PBA learns variance; avoid manual budget edits | 
| Stabilisation | Weeks 2–3 | Monitor Smartly Budget Movement logs and Fospha KPIs | 
| Evaluation | Weeks 4–5 | Compare test vs control PoP in Fospha | 
| Scale/Pause | Weeks 6–8 | If results ≥10% uplift, scale to more markets/channels; if not, refine guardrails or extend test | 
Expect PBA to adjust budgets every ~2 days on Meta/TikTok and every 3–4 days on Pinterest/Snap.
5. Measuring Test Results
- Control design: Run a comparable campaign/market outside the pool on platform default optimisation. 
- Compare Test vs Control in Fospha: - Use PoP to track CPA, ROAS, new-customer share. 
- Validate if budget shifts matched incremental outcomes. 
 
- Directional validation: Look for divergence in trend (e.g. falling CPA in test vs flat control). 
Success criteria:
- Look out for ≥10% CPA reduction or ≥10% ROAS uplift vs control. 
6. What To Do Next After the Pilot
Once the pilot completes, the goal is to decide whether to scale, refine, or re-run the test based on performance, learnings, and operational impact.
If Results Are Positive
- Scale gradually. - Add more campaigns into the existing pool or replicate the setup in adjacent markets with similar structure and KPIs. 
- Expand scope. - Move from single-channel to cross-channel pools once stable results are proven. 
- Build funnel-based pools. - Separate Awareness, Consideration, and Conversion pools to enable more granular control heading into peak seasons. 
- Layer in creative and bid optimisation. - Once PBA is stable, introduce creative tests or audience variations to compound gains. 
- Document uplift. - Capture the quantitative impact (≥10 % ROAS uplift or ≥10 % CPA reduction) and qualitative learnings (e.g. improved channel mix, reduced manual edits). 
If Results Are Neutral or Negative
- Extend the learning window. - Keep the test running for 6–8 weeks to allow the model to stabilise and gather more signal. 
- Re-check setup integrity. - Was the KPI consistent across campaigns? 
- Were creatives comparable? 
- Was there enough daily volume for learning? 
 
- Adjust guardrails. - Increase exploration budget (e.g. 10 → 15 %), widen min/max constraints, or slow the optimisation cadence to reduce volatility. 
- Simplify scope. - Split large pools into smaller, clearer tests (e.g. per channel or market) to isolate issues. 
- Re-baseline expectations. - Performance stabilisation may take longer for smaller budgets or upper-funnel campaigns; focus on directional improvement first. 

