Your Core Processor Is a Time Capsule — And That's Actually Your Biggest Asset
Everyone tells you your core is your biggest liability. Here's my contrarian take: the decades of data inside it are about to become your most valuable strategic asset.
Sean Hsieh
Founder & CEO, Runline
Article 5 Outline: "Your Core Processor Is a Time Capsule — And That's Actually Your Biggest Asset"
Track: The Market Thesis (deep blue) | Arc: Market Thesis Target: CEOs, CTOs Length: ~2,200 words
Opening Hook
Everyone tells you your core processor is your biggest liability. It's old. It's slow. It's written in a language nobody teaches anymore. Your vendor's "modernization roadmap" has been three years away for the last fifteen years. Here's my contrarian take: your core isn't a liability — it's a time capsule. And the decades of data trapped inside it are about to become your most valuable strategic asset. The question isn't whether to replace your core. It's whether to unlock the goldmine sitting inside it.
Act 1 — What's Actually Inside the Time Capsule
- 43% of banking systems worldwide still run COBOL. $3 trillion in daily commerce flows through these systems. This isn't a bug — it's a testament to how reliable they are.
- Your credit union's core processor — whether it's Jack Henry Symitar (built on IBM AS/400, RPG language, dating to the 1980s), Fiserv DNA, CU*Answers GOLD (56 years of continuous operation on IBM i-Series), or one of the newer entries like Corelation KeyStone — holds something no AI vendor can replicate: your institutional memory
- What's in there:
- 20-30 years of transaction history per member — every deposit, withdrawal, transfer, loan payment, fee, and reversal
- Loan origination and servicing records — complete lifecycle from application through payoff
- Member behavior patterns — channel usage, product adoption, seasonal patterns, life event indicators (address changes, beneficiary updates, new account types)
- Compliance records — BSA/AML alerts, CTR filings, SAR history, OFAC screening results, 7+ years of audit trails
- Communication history — every call, every dispute, every complaint, every resolution
- The problem isn't the data — it's the access. This data sits in proprietary formats, batch-processing cycles, flat files, and schemas that evolved through 56 years of append-only additions. Nobody designed it to be queried by AI. Nobody designed it to be queried by anything modern.
- Reference: I've been inside a CU*Answers data center. Their IBM Power server — $5M machine, 75 CPUs (only 13 in use) — runs a single LPAR (Logical Partition) with thousands of programs ranging from 500 to 40,000+ lines each. The schema metadata? "Not strong." Transaction categorization? Seven generic buckets instead of granular merchant category codes. The MCC codes are in the data — they're just not surfaced to consumers or analytics. The data is there. The access layer isn't.
Act 2 — Why "Rip and Replace" Is the Wrong Answer
- The CU industry has been talking about core modernization for 20 years. The track record is... not great.
- Core conversions take 12-24 months, cost millions, and carry catastrophic risk. One failed conversion can mean weeks of member-facing outages, regulatory scrutiny, and permanent trust damage
- Jack Henry collected $16M in deconversion fees in FY2025 alone — that's the penalty for leaving, before you've spent a dollar on the new system
- 69% of financial institutions plan to stay on their current core (American Bankers Association). They're not staying because they love it — they're staying because the switching cost is existential
- The cloud-native alternatives (Thought Machine, Mambu, Temenos) are real and improving, but they're designed for neo-banks and fintechs, not for institutions with 30 years of member history, complex product configurations, and regulatory obligations that require data continuity
- The dirty secret: Even when a CU does convert, they often lose historical data in the migration. Legacy formats don't map cleanly to modern schemas. Decades of institutional memory — gone.
- So here's the real question: If 69% of CUs aren't replacing their core, and even the ones that do risk losing their historical data... what if the answer isn't replacing the core at all? What if it's building an intelligence layer on top of it?
Act 3 — The Data Layer Thesis (Lessons from the Giants)
- The most valuable companies in technology share one trait: they built an access and intelligence layer on top of data that already existed
- Bloomberg didn't create financial markets. They built a terminal that normalized, indexed, and made accessible the data that was already flowing through trading desks. Revenue: $10B+/year from what is fundamentally a data normalization and access product.
- Palantir didn't create government intelligence data. They built Gotham — a platform that integrated, normalized, and made queryable the data that was already sitting in disconnected government databases. The data existed. The access layer didn't.
- Plaid didn't create bank accounts. They built an API layer that made consumer financial data accessible to applications. They're now the data layer for fintech.
- The pattern: The data already exists. The value is in making it accessible, normalized, and intelligent. Nobody needs to create new CU data. Your 30 years of member history, transaction patterns, and institutional knowledge are already there. You need a way to unlock it.
- Reference: Morgan Stanley gave their advisors RAG-powered (Retrieval Augmented Generation) access to 100K+ internal documents. Adoption hit 98%. The documents already existed. The AI access layer made them useful for the first time. The same pattern applies to your core processor data.
Act 4 — CDC: Unlocking the Time Capsule Without Breaking It
- The technology to do this exists today. It's called Change Data Capture (CDC) — and it's how you build a real-time intelligence layer on top of your legacy core without touching, replacing, or risking it.
- How it works: CDC reads the core processor's internal journal (every database change is already logged by the IBM i-Series for system integrity). A CDC engine — like the open-source Debezium connector for IBM i — captures these changes in real-time and streams them to a modern data platform.
- Source: IBM i-Series journal → CDC Engine: Debezium → Event Bus: Apache Kafka → Storage: Parquet/ClickHouse → AI Layer: Semantic indexing, vector search, agent access
- What this means practically:
- Every transaction, account change, and member update is captured in real-time (sub-5 second latency), not nightly batch
- The data is normalized from proprietary formats into a modern, queryable schema
- AI agents can access the meaning of 30 years of member history without ever touching the core
- The core processor keeps running exactly as it does today. Zero risk. Zero disruption.
- The market comparison: IBM's proprietary CDC solution runs $600K-$1.5M over 3 years. Consulting firms charge $500K-$1M+ for data warehouse implementations. The industry standard timeline is 9-12 months. This can be done in 10-12 weeks with modern open-source tooling at a fraction of the cost.
- Reference: There is no cross-core data standard in the CU industry. CUFX has limited adoption. FDX is consumer-focused. This means every core integration is custom engineering work — understanding different APIs, data formats, journal structures, and decades-old schema decisions. This is hard, unglamorous work. It's also the most defensible moat in CU technology.
Act 5 — What Happens When the Time Capsule Opens (Closing)
- Once your core data is normalized, indexed, and AI-accessible, everything changes:
- Your BSA analyst stops manually querying three systems to investigate an alert. An AI agent pulls the member's full transaction history, account relationships, and prior alert patterns in seconds (Article 6 goes deep on this)
- Your loan officers see AI-generated member insights — predicted needs, risk signals, cross-sell opportunities — drawn from decades of behavioral data, not a single credit score
- Your CSRs get a complete member picture on every call — transaction context, communication history, product usage, life events — without toggling between 7 systems
- Your compliance team gets real-time monitoring, not nightly batch reports. Suspicious patterns detected as they happen, not 24 hours later
- Your examiners see a credit union with a documented, auditable, real-time data infrastructure — the kind of institution they hold up as a model, not a risk
- CU data utilization today: The average credit union scores 241 out of 500 on Cornerstone Advisors' data utilization index (2025). That means CUs are using less than half the value of the data they already own. The other half is sitting in the time capsule, waiting.
- Callback to Article 4: The SaaSPocalypse wipes out execution-layer vendors. But the data layer appreciates. Your core processor data — the same data everyone tells you is a liability — is the one asset that gets more valuable as AI gets more capable. Every other vendor in your stack is replaceable. Your 30 years of member history is not.
- Closing line direction: "Your core processor isn't old. It's seasoned. And in an AI era where data is the only durable competitive advantage, that 30-year time capsule is worth more than every modern SaaS tool in your stack combined. The question isn't whether to replace it — it's whether to finally unlock what's inside."
Key References
- 43% of banking systems run COBOL; $3T daily commerce flows through them
- IBM Power server architecture — $5M machine, 75 CPUs, LPAR partitioning
- CU*Answers GOLD — 56 years continuous operation on IBM i-Series
- Jack Henry — $16M in deconversion fees collected FY2025
- 69% of institutions staying on current cores (ABA)
- Core conversions: 12-24 month timelines, millions in cost
- Bloomberg — $10B+/year from data normalization and access
- Palantir Gotham — intelligence integration layer
- Plaid — API layer becoming fintech's data infrastructure
- Morgan Stanley — 98% adoption of RAG-powered internal document access
- Debezium — open-source CDC connector for IBM i-Series
- CDC market: IBM proprietary $600K-$1.5M (3-yr) vs. open-source at fraction
- No cross-core data standard: CUFX limited adoption, FDX consumer-focused
- Cornerstone Advisors — CU data utilization score 241/500 (2025)
- CU*Answers schema: 7 generic transaction categories, MCC codes exist but not surfaced
Tone Calibration
- Empathy: "I know you're tired of vendors telling you your core is broken. It's not broken — it's been running for decades, serving your members reliably every single day. The problem isn't the core. The problem is that nobody built the bridge between your core's data and the AI tools that could make it useful."
- Curiosity: Genuinely fascinated by what's inside these systems. The fact that CU*Answers GOLD traces back to Symitar/Episys lineage means structural patterns are shared across cores — which means normalization is hard but not impossible. The archaeology of CU data is an underappreciated intellectual challenge.
- Silicon Valley lesson: Bloomberg built a $10B/year business by making existing financial data accessible. Plaid built a multi-billion-dollar company by making existing bank data accessible. The pattern is clear: the data exists, the access layer doesn't. Credit unions are sitting on the same opportunity.
- Spicy take: "Everyone's selling you a new core. Nobody's offering to unlock the one you already have. That tells you more about vendor incentives than it does about your technology."
- Technical credibility: This article needs to show that Sean has been in the data center, explored the schema, understands the i-Series architecture. Not theoretical — experiential. The CU*Answers details (single LPAR, 7 transaction categories, MCC codes hidden in the data) demonstrate a depth that no slide deck can fake.