
Stabilizing a Fraud Detection System Under Admissions Pressure
An admissions platform began experiencing a sharp increase in fraudulent applications exploiting university-issued .edu email domains.
InHouse AI
12/18/2025
An admissions platform began experiencing a sharp increase in fraudulent applications exploiting university-issued .edu email domains. The system processed applications at high volume, with seasonal spikes that left little room for manual review. Admissions teams depended on fraud detection program to triage risk, but the existing approach had grown organically and lacked clear operational boundaries.
The organization was not experimenting. The system was already in daily use, and failures had real consequences: overwhelmed reviewers, delayed decisions, and erosion of trust in the platform’s screening signals.
The Risk
The primary risk was not model accuracy in isolation. It was operational fragility:
False negatives allowed fraud through at scale.
False positives created review backlogs and fairness concerns.
No clear ownership existed for failure handling or threshold changes.
During peak cycles, small degradations cascaded into large delays.
The system “worked”, but it was not dependable.
The Constraint
Several constraints shaped the work:
Time pressure: Admissions cycles left narrow windows to intervene.
Data reality: Labels were noisy and delayed; ground truth lagged decisions.
Human cost: Reviewers were already overloaded; new workflows had to be minimal.
Change tolerance: The organization could not pause operations for a rebuild.
A full re-architecture was unrealistic. The system needed to be stabilized in place.
The Decision
We evaluated multiple options:
Replacing the model with a more complex architecture
Introducing aggressive automation to maximize catch rates
Rebuilding the pipeline end-to-end
We rejected these paths. Each increased complexity, delay and more risk, without addressing the core problem: lack of control and clarity in production.
Instead, we focused on:
Defining explicit decision boundaries and confidence thresholds
Separating detection from action (signal vs. workflow)
Introducing conservative fail-safes during peak load
Establishing ownership for monitoring, overrides, and drift review
The goal was not to catch everything. It was to make the system safe to rely on.
Outcome
After stabilization:
Approximately 95% of fraudulent applications were flagged before review
Manual review effort dropped by ~50% during peak periods
Reviewers started to trust the system’s signals and escalations once its results were found to be reliable.
Admissions decisions became faster and more consistent
Just as important, the organization gained clarity on what the system could and could not be trusted to do.
What This Taught Us
Accuracy is not reliability. Systems fail operationally long before they fail statistically.
Boundaries matter more than models. Clear thresholds and ownership reduce risk faster than complexity.
Stabilization beats reinvention. In production environments, control often delivers more value than novelty.
AI systems that influence real decisions must be designed for dependability first. If a system cannot be trusted during peak pressure, it is not ready for daily use.
What we do


AI, from Plug-In to Built-In
Company
Contact
