How inference works
Last updated
Last updated
The Airflux inference engine is an AI-powered system that makes real-time decisions on whether to display an ad or not at a particular moment for maximum ad revenue. When a player reaches a point where an ad is supposed to be displayed, the Airflux SDK sends the in-game events and player attribute data to the Inference server. Using the data, a decision regarding ad display is made and returned to the client. The inferencing process includes the following steps:
Data Collection
The SDK gathers data such as in-game events, player attributes, revenue, device information, and other relevant data in real-time.
Data Preprocessing
The collected data is converted into formats usable by the AI engine, including recent event count, session information, and eCPM metrics.
Model Inference
The AI engine selects a model based on player profile and game context, and performs inference to determine optimal ad display timing.
Ad Display Decision
The AI engine returns a result aligned with the optimization goal, indicating whether or not to display an ad.
The Airflux AI engine takes into account a variety of variables. The variables detailed in the table below are used to identify the best timing and approach for displaying ads, maximizing ad revenue while ensuring a positive play experience.
Player profile
Country, OS, Language
Player segmentation
In-session behavior
Accumulated playtime, recent session duration, event count
Churn and retention prediction
Ad response
Recent responses to ads
Ad resistance assessment
Revenue signal
The recent average of in-app revenue amount, eCPM
High-value user identification
Game context
Current stage, currency inventory
Adjustment for optimal play experience
Time series signal
Current date, day of week
Optimization around holidays and weekends
The models used by the Airflux AI engine evolve through several training loops:
Online feedback loop: Internal model parameters are adjusted based on ad responses to each API call. Underperforming policies are discarded.
Model retraining loop: When a large volume of events is aggregated or significant pattern changes are detected (e.g., new stages, patches), retraining is triggered.
Version update: The data science team continues to develop and deploy new, improved models that outperform the previous versions.
The Airflux inference engine’s architecture is built on the following key components:
CloudFront Global CDN: All inference requests are routed through CloudFront, a top-tier global CDN, to the nearest PoP (Point of Presence), minimizing latency.
AWS Services: Airflux uses services such as Application Load Balancer, Lambda@Edge, Aurora (RDS), DynamoDB, and S3 to ensure high durability and scalability across multiple Availability Zones (AZs).