Step 7. Score for what you can actually build

Step 8 of 9

Step 7. Score for what you can actually build

You now have a data map and a list of use cases scored for business impact. The second scoring dimension is feasibility: how hard is it to actually build this?

Feasibility has two components, and you should score them separately.

Data feasibility — does the data exist, is it accessible, and is it good enough? Use your data map from Step 6 to assess each use case. A use case that requires attributes you know are missing or unreliable will score low here regardless of how valuable it would be.

Capability feasibility — does your team have the skills to build it? A churn prediction model requires different skills than a dashboard. A real-time alerting system requires different infrastructure than a monthly report. Be honest about what your team can deliver in the near term.

Score each dimension on a 1–5 scale:

ScoreLabelDescription
5Ready nowData exists, is accessible, clean, and has sufficient history. You could start building tomorrow.
4Minor gapsMostly there but needs some cleaning, joining, or enrichment. A few weeks of work.
3Achievable with effortData exists but is fragmented, poor quality, or hard to access. A meaningful engineering project.
2Significant investmentKey data is missing or requires new capture mechanisms. A multi-month effort to get ready.
1Not currently possibleData does not exist and cannot be readily created. Requires new processes or systems.

The overall feasibility score for a use case is the lower of the two component scores. A use case that is a 5 on data but a 2 on capability is still a 2 overall — you cannot build what you do not have the skills for.

Keep the component scores visible in your working. A use case that is a 5 on data and a 2 on capability is a hiring or partnering problem. One that is a 2 on data and a 5 on capability is a data engineering problem. These have different solutions and different timelines.

Example — B2B SaaS

StageQuestionDataCapabilityOverallNotes
MarketingWhich campaigns are driving signups, and what is the cost per trial?444Marketing and product data need joining, but both exist and are accessible.
MarketingWhich campaigns are most likely to bring in accounts that convert and stay?333Requires joining marketing source data to long-run retention outcomes, which needs history.
SalesWhat is our win rate by deal size and segment? Where are we losing?444CRM data is typically clean enough for this.
SalesWhich prospects are most likely to close, and which should the team prioritise this week?322Needs engineered features from CRM activity. ML model required.
SignupWhere are people dropping off in the signup flow?544Product analytics tools capture this natively.
SignupWhich visitors are most likely to complete signup if we intervene?322Propensity model requires behavioural features and ML capability.
OnboardingWhat proportion of new accounts complete onboarding, and how long does it take?444Product events are typically well-instrumented for this.
OnboardingWhich new accounts are struggling and need intervention before they disengage?433Data is available; threshold-based alerting is achievable without advanced ML.
ActivationWhat is our time to first value, and how does it vary by segment?444Standard product analytics work once activation event is defined.
ActivationWhich users are at risk of never activating, and what nudge is most likely to help?322Recommendation model for nudge type adds significant complexity.
Ongoing usageWhich features are being used and which are being ignored?544Product event data is typically the most mature data asset in a SaaS business.
Ongoing usageWhich product investments would most improve retention? Which disengaged users are about to churn?322Feature importance analysis on retention requires clean cohort history; churn model adds ML requirement.
ExpansionWhich accounts have grown usage without upgrading their plan?444Usage and billing data join is straightforward.
ExpansionWhich accounts are ready to expand, and what is the right offer and timing?322Propensity model on usage signals plus offer recommendation adds two layers of complexity.
RenewalWhat is our renewal rate by segment and cohort?444Billing data is typically accessible.
RenewalWhich accounts are at risk of not renewing, and what should we do about each one?322Churn model with next-best-action recommendation requires ML engineering.
ChurnWhy are customers leaving, and does the reason vary by segment?333Needs exit survey data combined with usage history — data quality varies significantly.
ChurnWhich active accounts are most likely to churn in the next 90 days?322Binary classification model on usage and engagement features.
Win-backWhat is our win-back rate, and which types of former customers return?444Historical billing and CRM data is sufficient.
Win-backWhich churned accounts are worth pursuing, and what offer has the best chance of working?322Win-back propensity plus offer optimisation — two model layers.

Example — Non-bank lender

StageQuestionDataCapabilityOverallNotes
MarketingWhich campaigns are generating applications that go on to settle? What is the cost per settled loan?343Marketing platform and application system need to be joined — UTM tracking through to settlement is often missing or inconsistent.
MarketingWhere should we shift budget to get more of the right applications?333Needs settled-loan attribution working first, plus a channel optimisation model.
ReferralWhich brokers are sending the most volume and the best quality applications?444Broker data and application outcomes are typically well-captured.
ReferralWhich brokers should we invest more in, and which are costing more than they are worth?433Data is available; commission cost-benefit model requires some engineering.
ApplicationWhat is application volume by channel and how is the mix trending?544Core operational data, typically clean and accessible.
ApplicationAre there sources of applications we are underinvesting in?333Requires market-level benchmarks as well as internal data.
Pre-approvalWhat proportion proceed to full approval, and where do they fall over?444Stage-level conversion data is standard in the application system.
Pre-approvalWhich applications are likely to fail at approval so we can flag them early?322Predictive model on application features requires ML and sufficient historical decline data.
Document collectionHow long is this taking, and where are the delays?444Timestamps on document receipt events are usually captured.
Document collectionWhich applications are going to miss settlement targets unless we intervene?322Time-to-completion model requires historical stage timing data and ML capability.
ApprovalWhat is our approval rate, and how does credit performance compare to what we expected?433Approval data is accessible; comparing predicted vs actual default performance requires model monitoring capability.
ApprovalAre there applications in the queue we should prioritise or escalate?422Queue prioritisation scoring model adds ML complexity on top of accessible data.
Settlement and fundingWhat proportion of settlements are on time, and where do delays come from?444Settlement timestamps are core operational data.
Settlement and fundingWhich loans in the pipeline are at risk of not settling on time?322Predictive model on pipeline features.
ServicingHow is the book performing by cohort, product, and channel?544Servicing platform data is typically rich and well-structured.
ServicingWhich borrowers are showing early signs of stress before they miss a payment?322Early warning model on behavioural signals requires feature engineering and ML.
HardshipWhat proportion of hardship arrangements result in the borrower returning to normal repayments?444Outcome tracking on hardship cases is standard in case management systems.
HardshipWhich borrowers are at risk of entering hardship in the next three months?222Limited behavioural signals available before hardship is declared; model requires data that is often not captured.
ArrearsWhich contact approaches are most effective at resolving arrears?333Requires contact activity logged against outcomes — data quality varies by collections team.
ArrearsWhich borrowers in arrears are most likely to resolve without escalation, and which need immediate action?322Resolution prediction model on arrears history and borrower features.
Discharge or refinanceHow many customers are leaving at refinance and what is the typical timing?444Discharge data is typically well-captured.
Discharge or refinanceWhich borrowers are most likely to refinance away in the next six months, and is there an offer that would retain them?322Prepayment prediction plus retention offer optimisation — two model layers.

If your capability scores are dragging down use cases you know are worth building, that is exactly the problem our platform is designed to solve. TrueState gives teams without large ML engineering headcounts the ability to build the predictive and prescriptive use cases that would otherwise sit in the strategic bet column indefinitely.

For teams that have completed this strategy process, we offer free onboarding — usually priced at $35,000. Book a walkthrough to see how it works.