15 December 2025

Propensity modelling in the age of consent

Bees working

In summary

  • What: Propensity and predictive models are being reshaped by consent, minimisation and purpose limitation, changing how data can be collected and used.
  • Why: Privacy rules restrict the individual-level, historical data these models rely on, and consent states now shift dynamically user by user.
  • When: As third-party cookies disappear and global privacy reforms accelerate, especially in Australia, consented, purpose-led data becomes the only durable modelling foundation.
  • How: By embedding privacy into data architecture through pseudonymisation, purpose-led pipelines, aggregated or cohort-based modelling, strong data-layer design, and awareness of implicative data risks.

For years, marketing analytics operated on a simple mantra: collect everything, analyse later. More data meant better models, better targeting, and better growth.

But the industry has crossed a threshold. Global privacy laws (GDPR, CCPA and incoming reforms in Australia), rising user expectations, and structural changes like third-party cookie deprecation have shifted the ground beneath every data-driven discipline.

Today, the success of any model, whether it’s propensity, LTV, recommendation systems, time series forecasting, MMM, or MTA, hinges on something far more foundational than statistical power:

Do you have consent to use the data your model is learning from? Welcome to the age of permission-led modelling.

Historically, brands treated consent as a simple yes/no toggle. But modern regulations require something much more nuanced:

  • Consent must be specific to a purpose
  • Consent must be informed and unambiguous
  • Consent can be withdrawn at any time
  • Consent does not automatically carry over to analytics or advertising

This means every user sits in a multi-faceted consent state, and your models must adapt in real time.

Challenge 1: Purpose limitation

If a user gives data to fulfil an order, that does not grant permission to use it for marketing analytics.

This one rule alone disrupts almost every legacy data architecture.

Impact on modelling

Models can no longer be trained on one unified customer table.

Every data field needs to carry metadata about:

  • Original purpose
  • Legal basis
  • Allowed downstream uses

Example:

  • Sales data collected under Contract
  • Marketing attribution requires Consent
  • You cannot simply join these datasets unless the user explicitly agreed.

This breaks the long-standing industry habit of “just add that column to the model”.

Most consent banners now offer categories:

  • Analytics
  • Personalisation
  • Advertising
  • Cross-site tracking

This means:

Your training dataset is no longer “all users for all purposes,” it’s dynamically filtered based on the permissions each user actually granted.

Legacy models assumed stable, universal data coverage. Today’s privacy environment breaks that assumption entirely.

Impact on common marketing models

1. LTV and propensity models

These models depend on deep historical data, exactly the data most vulnerable to minimisation rules and deletion requests.

The problem:

  • Minimisation reduces the depth of data you’re allowed to hold
  • Purpose Limitation fragments your training cohort

Outcome: Lower predictive accuracy and unstable features.

The solution: Pseudonymised training pipelines

Instead of training on identifiable user data, brands should:

  • Strip direct identifiers into a secured PII table
  • Assign a pseudonymised ID
  • Train models on behavioural patterns, not individuals
  • Invalidate IDs when users delete their data

This preserves model utility while honouring legal obligations.

2. Marketing mix modelling (MMM)

MMM is aggregated by nature, but it still relies on identifiers and conversion logs.

The problem:

Cookie deprecation and opt-outs create “holes” in the dataset.

The solution: Synthetic data generation

Where user-level data is missing for privacy reasons, synthetic datasets, built from aggregated patterns, can fill statistical gaps without re-identifying individuals.

3. Multi-touch attribution (MTA)

MTA tries to stitch impressions = clicks = conversions across long timelines.

This is directly at odds with modern privacy obligations.

The problem:

Persistent identifiers are fading or illegal without ongoing consent.

The solution: Move to cohort-based or differential privacy MTA

Focus on group behaviours and aggregate flows. Not individuals.

This shift mirrors what Google and Meta are doing with privacy sandboxes and aggregated event measurement.

How privacy-first data architecture solves these challenges

Privacy-compliant modelling isn’t just a legal exercise, it’s a structural one. Below are the core architectural patterns Louder regularly recommends.

A dedicated Consent Service tracks:

  • User ID
  • Consent type
  • Legal basis
  • Timestamp
  • Expiry

Every analytics pipeline must check permissions before processing data.

This transforms consent from a front-end banner into an enforceable rule baked into your data model.

2. Architectural Separation (Pseudonymisation)

Split your system into:

  • PII Table (minimised, restricted, highly privileged access)
  • Behavioural Data Table (rich, useful, non-identifiable)

Link them only through a pseudonymised ID stored in a secure linkage table.

Why it matters:

When a user invokes Right to Delete:

  • You delete one row from the PII table
  • The pseudonymised ID becomes invalid
  • All behavioural history becomes unlinked and therefore unusable

This is a privacy-safe kill switch.

3. Embedded data lineage, retention and auditability

Every data field should include:

  • Its consent state
  • Its legal basis
  • Its retention period
  • Its source system

Your models should trigger automatic deletion or anonymisation when:

  • Consent expires
  • Purpose changes
  • Retention limits are reached

This is where data layers become a key component. A well-structured data layer ensures:

  • Only the right data is collected
  • Every field is tied to a purpose
  • Downstream systems know how it can be used

Where this leaves the industry

The age of consent doesn’t mean the end of modelling. It means the end of lazy modelling.

The organisations that win will be the ones that:

  • Treat consent as a dynamic model input
  • Build resilient architectures that survive data deletion
  • Shift from individuals to pseudonymous or cohort-based signals
  • Prioritise declared and contextual data over passive surveillance
  • Use synthetic data where privacy removes granularity
  • Invest in privacy engineering as a core analytics capability

Prediction must now be grounded in permission.

Louder’s recommendations

  • Build privacy into the data model itself. Consent must be machine-readable and enforced automatically across every pipeline.
  • Use purpose-led, pseudonymised data architectures.Models should never access raw identifiers, only the minimum data required for the approved purpose.
  • Shift from individual-level to aggregated or cohort-based modelling where needed. A more privacy-resilient approach for attribution, measurement and forecasting.
  • Treat the data layer as critical infrastructure. It ensures only the right data is collected, with clear purpose, provenance and retention.
  • Use synthetic or cohort-based methods to offset signal loss. Future-proof models by designing them to operate even when granularity drops.
  • Train teams in “implicative data” risk. Privacy risk increasingly comes from what models can infer, not just what’s collected.

Get in touch

Get in touch with Louder to discuss how we can assist you or your business and sign up to our newsletter to receive the latest industry updates straight in your inbox.



About Archit Sharma

Archit is an analytics consultant at Louder with a background in marketing, stats and programming. In his spare time, he enjoys playing football, curating playlists, or trying out a new single malt.