AI lead scoring in mortgage is one of the few production-grade AI applications that consistently produces measurable ROI for loan officers and mortgage businesses. Done right, it lifts closed-loan rate 15-25% on the same lead volume by changing which leads get worked first. Done wrong, it creates a false sense of priority while missing the actual conversion drivers.
This piece breaks down what the model actually does, the data it uses, what good performance looks like, and where the limits are.
What is a propensity model in mortgage?
A propensity model is a statistical or machine-learning model that estimates the probability of a specific outcome. In mortgage lead scoring, the outcome is “this lead will result in a funded loan within X days.” The model takes a lead’s attributes and behavioral signals as inputs and outputs a probability score, typically 0 to 100 or 0 to 1.
The model is trained on the mortgage company’s own historical data: leads that closed, leads that did not, and the attributes of each. Patterns the model learns from this data become the rules it applies to new leads. A lead with attributes similar to leads that historically closed gets a high score; one with attributes similar to leads that did not gets a low score.
What data does the model use?
Production mortgage AI lead scoring models use a combination of three data categories:
Borrower attribute data
- Stated income range
- Loan amount range
- Property type and location
- Loan purpose (purchase, refinance, cash-out)
- Borrower stated credit range
- Number of properties owned
- Self-employment status
Behavioral signals
- Response time to first LO outreach
- Email opens and clicks during nurture
- Document upload speed and completeness
- Calendar booking behavior
- Number of return visits to the LO website
- Application form abandonment vs completion
Source and contextual data
- Lead source (organic, paid, referral, partner)
- Marketing campaign attribution
- Day of week and time of day of inquiry
- Current market rate environment
- Local market conditions
The model assigns weights to each input based on its predictive power. For typical mortgage models, the strongest single predictor is response time to first outreach (which is why the 5-minute rule matters so much). Second is lead source attribution. Third is borrower-stated income relative to typical loan size in the geography.
What does good lead-scoring performance look like?
Useful production models hit 70-80% accuracy on the top-decile predictions. That is, of the leads the model scores in the top 10% of probability, 70-80% close within the prediction window. Below that threshold the model is not adding signal beyond what the LO would intuit from the lead data alone.
| Score band | Typical close rate | What LOs do |
|---|---|---|
| Top 10% (highest scores) | 30-50% | Call within first 30 minutes |
| 11-30% band | 10-20% | Call within first 2 hours |
| 31-60% band | 4-8% | Email-led nurture, call within 24h |
| 61-100% band | 1-3% | Long-tail email nurture only |
The point of the score is not to write off the bottom band. Some of those leads still close. The point is to allocate LO time disproportionately to the top band where each call has 10-20x the conversion probability.
How long does it take to build a useful model?
Production-grade lead scoring requires:
- Minimum 200-500 historical funded loans for training
- Equal or greater number of leads that did not close (negative class)
- 12+ months of behavioral data on those leads
- Clean attribution between lead source and closed-loan outcome
For most active mortgage operations, this data exists in the LOS plus the CRM, but it is rarely cleaned and matched. The first 30-60 days of a lead-scoring implementation goes into data preparation, not model training.
For brand-new operations or LOs without sufficient history, lead scoring is not feasible. The alternative is rule-based scoring (response time + lead source + stated loan amount) which is less accurate but does not require historical training data.
Where does AI lead scoring fall short?
- Market regime changes. Models trained in low-rate environments produce different predictions than models trained in high-rate environments. Quarterly retraining is required to keep the model current.
- New lead source types. When a new lead source goes live (a new ad platform, a new partner channel), the model has no training data for that source until enough leads close from it. Allow 30-60 days before trusting scores on new sources.
- Borrower self-reported data accuracy. Models depend heavily on borrower-stated attributes (income, credit, loan amount). Borrowers misstate or exaggerate these. Cross-reference against verified data when possible.
- Compliance constraints. Lead scoring models cannot use protected-class attributes (race, ethnicity, age, gender, marital status, religion, disability) or proxies for them. ECOA and HMDA compliance limit what features the model can include.
For most mortgage operations doing 30+ loans per month with clean data, AI lead scoring is one of the highest-ROI tech investments available in 2026. For smaller operations or those with inconsistent data, simpler rule-based scoring delivers most of the benefit without the implementation overhead.
Common questions
How accurate is mortgage AI lead scoring?
Production-grade models hit 70-80% accuracy on top-decile predictions. The top 10% of scored leads typically close at 30-50% rates, vs. 1-3% for the bottom band. Accuracy degrades when market conditions shift, requiring quarterly retraining.
Do I need a data scientist to use AI lead scoring?
No, if your CRM has it built in. Most modern mortgage CRMs (BNTouch included) ship with lead scoring that runs against your data automatically. Custom AI implementations require data science expertise; embedded CRM features do not.
How does lead scoring interact with fair-lending compliance?
Lead scoring models cannot use protected-class attributes (race, ethnicity, age, gender, marital status, religion, disability) or proxies for them. ECOA and HMDA compliance constrain what features the model can include. Compliant implementations are auditable and use only legitimate business factors.
Can I see the actual score for each lead?
Yes, in most CRMs that offer lead scoring. The score appears on the lead record and in the pipeline view. Some platforms also expose the top contributing factors so the LO understands why a lead is scored high or low.
Is AI lead scoring better than my LO’s gut feel?
On large volume, yes. Models process more variables than a human can hold in mind and detect patterns across thousands of historical loans. On individual cases, an experienced LO can sometimes spot factors the model misses. The right approach is using the model for triage and the LO for judgment.
See AI lead scoring on your data.
BNTouch’s lead scoring runs against your historical loan data. Free demo includes a sample model trained on representative data.



