The Sopact Intelligence Library
Book 06 · Industry Guide

Nonprofit
Programs.

Surveys, transcripts, partner reports, offline data — seven systems or one record. The 95% of stakeholder voice that nobody reads is the program intelligence you're missing.

— CHAPTER ONE
The Unread 95%.
Architecture · Multi-source · Multi-language · Continuous intelligence
DESIGN 01 Theory of Change from interviews COLLECT 02 Unified · Multi-source 40+ languages REPORT 03 Continuous Intel 6 reports · auto → ONE stakeholder_id ACROSS PARTNERS · PROGRAMS · COHORTS · YEARS →
A SOPACT INTELLIGENCE LIBRARY GUIDE · Vol. 06
1
CONTENTS
— TABLE OF CONTENTS

In this book.

Seven chapters on how nonprofit programs, foundations' grantees, and multi-program organizations move from seven systems to one connected record. Same architecture across partners, cohorts, languages, and years.

Book 06 · Chapters
01
The Unread 95%.
YOU ARE HERE
02
Theory of Change from Interviews
14 MIN
03
Multi-Source Collection — Online · Offline · Documents
13 MIN
04
The Multi-Language Pipeline
12 MIN
05
Qualitative Coding at Scale
15 MIN
06
Partner & Board Reporting
13 MIN
07
Cross-Program Patterns & Learning
11 MIN

Each chapter is published as a stand-alone PDF and as a sub-thread inside this book. The whole book reads in roughly ninety minutes. The chapters are self-contained.

The library
BOOK 01
Beyond the Survey
Foundations — the five-stage spine under everything else.
BOOK 02
Application Management
Reviewer workflows, scoring, fairness — any application pipeline.
BOOK 03
Grant Intelligence
Foundations, family offices, community grant-making.
BOOK 04
Impact Intelligence
Impact funds, ESG, supply chain DD, LP reporting.
BOOK 05
Training Intelligence
Workforce programs, accelerators, L&D, cohort outcomes.
BOOK 06 · CURRENT
Nonprofit Programs
Multi-program service delivery, partners, board & funder narrative.
2
CHAPTER 01
01
— BOOK 06 · CHAPTER ONE
NONPROFIT PROGRAMS

The Unread
95%.

The frameworks were never the problem. Of all the stakeholder voice your programs collect, roughly 95% sits in PDFs, transcripts, open-ended responses, and partner reports nobody ever reads. The funder asks "what changed?" The honest answer is in the 95% — and the dashboard is built from the 5%.

What you'll learn
  • Why every purpose-built impact-measurement platform from the past decade either shut down, pivoted to ESG, or stalled — and what they all got architecturally wrong.
  • How a Theory of Change gets built from interview transcripts you already have — not consultant workshops you can't afford.
  • The 80% cleanup tax that consumes most M&E staff time — and what changes when data arrives clean from collection.
  • How 1,000 qualitative responses get coded in four minutes — across 40+ languages, with citation trails.
Read time
15
minutes · 16 pages · ~22 visuals

Skill files referenced
toc-from-interview.md multi-language-coder.md partner-report-aggregator.md board-narrative-composer.md
3
§ 1.1 · WHY
— SECTION 1.1

Seven systems. Eighty percent cleanup.

The theory of change lives in a Google Doc. Survey data is in Airtable. Financial reports arrive as PDFs. Qualitative notes sit in Excel. Offline submissions arrived in KoboToolbox. Interview transcripts are in DOCX — half of them in Portuguese. Two different teams are translating the same field data.

Research is consistent on this. 76% of nonprofits say impact measurement is a priority. Only 29% are doing it effectively. The gap isn't ambition — it's that the field spent fifteen years building increasingly sophisticated frameworks on top of fundamentally broken data architectures, then blamed organizations for "lacking capacity" when they couldn't implement what the frameworks demanded.

— WHAT YOU COLLECT VS WHAT YOU ANALYZE
100% collected — surveys · transcripts · PDFs · offline · open-ends 5% analyzed — what fits a dashboard 95% UNREAD → 5% effort produces the report · 95% of stakeholder voice goes unanalyzed

The software market has already voted. Social Suite and Sametrica pivoted to ESG. Proof.io and iCuantix ceased operations. Impact Mapper retreated to a consulting model. SureImpact, UpMetrics, and ClearImpact require ongoing managed services to stay operational. When every purpose-built platform in a category either shuts down or retreats from software to services, that's not individual company failure — that's market failure.

They all made the same mistake — they started with frameworks and dashboards instead of solving the data architecture problem underneath. They asked "What metrics should we track?" when the real question was "How do we collect context that's actually usable?" — BOOK 06 · CH 01 · §1.1
4
§ 1.2 · BIG IDEA
— SECTION 1.2

Three phases. One stakeholder. Five colors.

The nonprofit-program lifecycle has three operational phases — program design, data collection, and continuous intelligence — and they map cleanly onto the five-stage spine that runs through every book in this library.

Read across: at design, Effective Data flows into a Theory-of-Change framework. By collection, the framework is anchored in a Data Dictionary tied to one persistent stakeholder ID. Every survey, transcript, document, and offline submission becomes a Transformation that feeds the Reports — six funder-ready outputs across partners, programs, cohorts, and years.

— PROGRAM PHASE × SPINE STAGE

S1 · DATA
Multi-source intake · online + offline + docs + transcripts
DESIGN + COLLECT
S2 · FRAMEWORK
Theory of Change · Logic Model · IMP 5D · IRIS+
PROGRAM DESIGN
S3 · DICTIONARY
Shared question library · persistent stakeholder_id · 40+ languages
DATA COLLECTION
S4 · TRANSFORM
AI codes qual + quant · themes across 1,000s of responses
CONTINUOUS INTEL
S5 · REPORTS
6 reports · partner dashboards · board narrative · any language
CONTINUOUS INTEL
id

One persistent stakeholder_id across partners, programs, cohorts, and years.

The participant who completes an intake form in January appears as "Maria Garcia" in one dataset, "M. Garcia" in another, "Maria G" in a third. Connecting these records is manual matching that introduces errors and restarts every cycle. Sopact Sense issues the ID at first contact — and every subsequent survey, document, partner report, or offline submission inherits it. The 80% cleanup tax doesn't go down. It disappears.

5
§ 1.3 · STAGE 1 · DATA
S1 · DATA — FOUR INPUT TYPES · ONE RECORD

What a real program actually receives.

Survey rows. Field notes in three languages. Annual partner reports as 100-page PDFs. Offline submissions in XML batches from KoboToolbox. Interview transcripts a coordinator translated and never had time to code. This is the actual shape of nonprofit data — and the 80% cleanup tax is where the field has been stuck.

Surveys · Online + Offline STRUCTURED

  • Pre / mid / post / 90-day surveys
  • KoboToolbox / ODK offline submissions
  • SurveyCTO · Qualtrics import
  • Unique reference links — one per stakeholder
  • Locked scales · paired by stakeholder_id

Documents UNSTRUCTURED

  • Annual partner reports (PDFs, 50–200 pages)
  • Financial statements, audits
  • Existing Theory of Change drafts
  • Funder-specific narrative templates
  • Read end-to-end · scored at upload

Interview Transcripts QUALITATIVE

  • Beneficiary interviews — any language
  • Coordinator field notes (Spanish, Portuguese, Swahili…)
  • Mentor session recordings
  • Partner check-in transcripts
  • AI-coded · linked to stakeholder_id

Partner Operational Data CONTEXT

  • School attendance records
  • Meal distribution logs
  • Case-management notes (Apricot, ETO)
  • Email attachments from partners
  • Read-only API pulls — partner stack unchanged

The
Cleanup
Tax

A failure pattern named.

80% of M&E staff time goes to cleaning, deduplicating, and reconciling data across Airtable, Excel, Google Forms, and KoboToolbox exports. By the time the data is clean, the program cycle is over. The cleanup tax is not a deficiency to be fixed in the team. It's an architecture choice made before the first survey went out — and the fix is upstream, not downstream.

6
§ 1.4 · STAGE 2 · FRAMEWORK
S2 · FRAMEWORK — THEORY OF CHANGE · FROM CONVERSATION, NOT WORKSHOP

Stop overthinking the framework. Start collecting.

Organizations spend months — sometimes years — designing the perfect Theory of Change, debating indicator definitions, hiring consultants to refine causal pathways. And then nothing happens. The framework sits in a PDF nobody opens. Data collection never starts.

The organizations that build real measurement capability share a pattern: they start collecting, not planning. With AI-native tools, the framework emerges from conversations already happening — calls between funders and grantees, coordinator interviews with field staff, coaching sessions with participants. The four frameworks below are reference points, not gates.

FRAMEWORK 01
Theory of Change Causal pathway from activities → intermediate outcomes → long-term impact. Best for program design.
FRAMEWORK 02
Logic Model Inputs → Activities → Outputs → Outcomes → Impact. Linear and practical for established programs.
FRAMEWORK 03
IMP 5 Dimensions What · Who · How Much · Contribution · Risk. Standard for impact-investor reporting.
FRAMEWORK 04
IRIS+ Metrics Catalog of standardized indicators from GIIN. Useful for benchmarking and peer comparison.

— LIVING TOC · GENERATED FROM A CONVERSATION YOU ALREADY HAD
Existing Conversation Funder ↔ grantee call · coordinator field interview · coaching session
+
Program Description Whatever exists — a one-pager, a grant application, a board memo
=
Living ToC + Dictionary Indicators mapped, partners aligned, ready before Wave 1

Framework-
First
Thinking

A failure pattern named.

Invest six months designing the perfect Theory of Change. Discover your data collection cannot support it. Application data lives in email attachments. Feedback sits in Google Forms. Interview notes stay in someone's head. The framework was beautiful. The data architecture destroyed it. The reverse order works: build the architecture first, let the framework emerge from the conversations already happening.

7
§ 1.5 · STAGE 3 · DICTIONARY
S3 · DICTIONARY — ONE ID. MANY PARTNERS. MANY YEARS.

The record that survives every partner, every cohort.

A nonprofit running four programs across fifteen partner sites in three countries has — in standard stacks — fifteen versions of the same participant. Persistent IDs at first contact eliminate that. The participant who enrolled at Partner A's intake in March is the same record at Partner B's referral in September is the same record at the funder's annual report in December.

— stakeholder_id: IKAYA-018 · YOUTH PROGRAM PARTICIPANT

MULTI-YEAR ARC
IN Mar '25 Partner A intake ID issued offline / Kobo M3 Month 3 Field interview · PT AI-coded themes in 4 min, not 3 mo Y1 Year 1 close Post-survey + outcome Funder report ready morning of close Y2 Year 2 Partner B referral Same ID, new program no re-introduction Y3 Year 3 Alumni longitudinal Cross-program impact visible in board deck 5y+ Year 5+ Cross-partner pattern "3 partners share the same barrier" — ONE stakeholder_id: IKAYA-018 — CARRIED ACROSS PARTNERS, PROGRAMS, LANGUAGES, YEARS —

— SHARED QUESTION LIBRARY · ABRIDGED

Field Source ToC Stage Cadence
stakeholder_id Issued at first intake form Persistent across partners + years
partner_id · program_id Auto-tagged at intake Inputs Cross-tabulation key
barrier_open_end Intake + mid-program · any language Activities + outputs AI-coded by theme, linked to ID
outcome_indicator_x Post + 90-day · IRIS+ mapped Outcomes Pre/post pair, locked scale
partner_report_themes Annual partner PDF (100+ pp) Long-term impact Read at upload, themed at scale
narrative_quote Interview transcript · field notes Contribution evidence Citation per stakeholder_id
8
§ 1.6 · STAGE 4 · TRANSFORM
S4 · TRANSFORM — 1,000 RESPONSES IN 4 MINUTES

From three months of consultant time to four minutes.

Below: a working qualitative theme analysis from a single program cycle. 1,247 open-ended responses across four languages, AI-coded against the program's Theory of Change, every theme citing the source quote. What used to be three months of NVivo coding ends up sitting in the funder report before lunch.

Open Play Foundation · Ikaya Youth Program · Year 1 Themes

program_id · IKAYA-Y1   ·   Partners: 4 sites ·   Languages: EN · IsiZulu · IsiXhosa · Afrikaans
4 min
1,247 RESPONSES CODED
Theme 1 peer support
N=412 · 33% Mentioned across all four sites. Sample quote (translated from IsiZulu): "When I struggled in week three, the older girls came back to check on me. That's why I didn't leave." Cross-tabulated with retention: peer-support mentions correlate with 23% higher completion.
+23%retention
Theme 2 transportation barrier
N=287 · 23% Concentrated at Partner C (rural site). Sample quote (Afrikaans): "Die taxi-geld is te veel — sometimes I have to choose between class and food." Surfaces a partner-specific intervention before the next cycle.
3 of 4sites
Theme 3 confidence inflection
N=198 · 16% Linked to specific curriculum modules. Sample quote: "I used to think these things were for other people. Now I know they're for me too." Pairs with quantitative confidence-rating data on the same stakeholder_id.
Wk 5–6peak
Theme 4 staff turnover
N=89 · 7% Mentioned at Partner B and Partner D. Partner B is in cohort 3 of staff transition. Pattern matches risk flag raised in Year 0 partner DDQ — surfaces as early warning, not as post-mortem.
2 of 4sites
Theme 5 cross-partner pattern
CROSS · N=156 Three partners surface the same childcare-gap barrier independently. Partner-level reports would never have shown this. The roll-up across the same indicator across stakeholder_ids surfaces a strategic intervention for the foundation.
PORTFOLIOsignal

Every theme is queryable: which stakeholders mentioned it, which partners over-index, which quantitative outcomes correlate with it. The board narrative writes itself from this layer — and so does the partner-specific feedback for next year's grant cycle.


— EVIDENCE ANALYSIS · TRADITIONAL VS UNIFIED

Quantitative survey numbers
92%
96%
Open-ended survey free text
8%
94%
Transcripts interviews · field notes
3%
92%
Partner PDFs 100+ page reports
5%
88%
SEVEN-SYSTEM STACK UNIFIED RECORD
9
§ 1.7 · STAGE 5 · REPORTS
S5 · REPORTS — SIX REPORTS · ANY LANGUAGE · CONTINUOUS

Continuous learning, not annual reporting.

The reports below are not dashboards. They are publication-ready outputs — board decks, funder packets, partner-specific feedback — synthesized from the connected record. Every claim cites the source response or document. Multi-language collection in, multi-language reporting out — same data, any language the funder needs.

01

Program Impact Report

Theory-of-Change-aligned narrative — what changed, who experienced it, contribution evidence, risks. Qual + quant unified per program.

TOC-ALIGNED · ANY LANGUAGE
02

Missing Data Alert

Which partners haven't submitted, which fields are incomplete, which beneficiary cohorts have gaps — the day data is due.

DAY DATA IS DUE
03

Outcome Variance Report

Programs tracking below committed outcomes, with root cause extracted from open-ends. Where dropout is, why confidence dipped.

BEFORE BOARD REVIEW
04

Qualitative Themes Report

AI-coded open-ended responses — sentiment, themes, beneficiary voice at scale. 1,000+ responses synthesized in 4 minutes.

AS RESPONSES ARRIVE
05

Early Warning Report

Enrollment dropout signals, participation anomalies, partner performance flags. Surfaced as data arrives — not at year-end.

CONTINUOUS · AUTOMATIC
06

Partner + Board Summary

Executive narrative for board deck, funder quarterly packet, and partner-specific feedback. Evidence-backed, generated in minutes.

READY IN ANY LANGUAGE

Continuous, not annual.

Most M&E reports are backward-looking, annual, and arrive after the program cycle has already moved on. When the architecture holds, real measurement informs decisions while there is still time to act. When mid-program data shows a partner struggling, the intervention happens that week — not in next year's annual report.

10
§ 1.8 · WORKED EXAMPLE
— SECTION 1.8

Catalyst Foundation · 4 programs.

A composite mid-size foundation running four programs across three countries through fifteen partner sites. Youth workforce, food security, early childhood, women's economic mobility. Here is what the same data architecture produces across very different program shapes.

4
PROGRAMS · 15 PARTNER SITES
12k
STAKEHOLDERS · 3 COUNTRIES
7
LANGUAGES · COLLECTION → REPORT
1d
FUNDER REPORT · WAS 3 WEEKS
1

Month 1 — Existing conversations become the ToC

Catalyst doesn't build a Theory of Change in a workshop. The M&E lead uploads three existing artifacts: the foundation's most recent strategy memo, two transcripts from partner check-in calls, and the prior year's narrative report. Sopact synthesizes a Living ToC mapped to IRIS+ indicators across all four programs. Both the foundation and partner sites sign off in week two.

2

Month 2 — Collection starts. stakeholder_id at first contact.

Partner A starts taking intakes via KoboToolbox offline; Partner B uses an online link in Portuguese; Partner C uploads PDFs of paper forms. Every record inherits the same ID architecture. No VLOOKUP. No matching by email. The 80% cleanup tax doesn't get smaller — it disappears.

3

Month 4 — Early warning fires from a transcript

A partner-site coordinator records a beneficiary interview in IsiZulu, uploads it. The transcript surfaces a dropout-risk pattern that matches the early-warning taxonomy. The foundation's program officer sees the alert the same day — months before it would have appeared in quantitative attendance data.

4

Month 8 — Cross-partner pattern surfaces

Three partners surface the same transportation barrier independently. No single partner-level report would have shown it. The Qualitative Themes Report — rolled up across all stakeholder_ids — flags it as a strategic intervention for the foundation's next funding cycle. Foundations use this to surface portfolio-level patterns from narrative reports the program officers couldn't humanly read.

5

Month 12 — Funder report ready the morning the cycle closes

The annual funder report — Portuguese for one funder, English for another, Spanish for a third — generates from the connected record. Board narrative includes the cross-partner transportation insight, the dropout-risk save, and three cohort-level outcomes traced to specific curriculum components. Three weeks of scramble compresses to one day of review.

11
§ 1.9 · GALLERY
— SECTION 1.9

Same architecture. Five nonprofit shapes.

A single-program direct-service nonprofit and a multi-region foundation with fifteen grantees share the same data architecture but pull on different fields. Here is what the same connected record looks like across five very different organizations.

12
§ 1.10 · SIBLING BOOKS
— SECTION 1.10

This chapter closes the library.

Book 06 is the last volume in the Sopact Intelligence Library — and the one most nonprofits should read alongside one of the others. The connected-record architecture runs sideways across every book. Same spine, different lifecycle, different field names.

— BOOK 01 · FOUNDATIONS

Beyond the Survey

The five-stage spine in full — Effective Data, Framework, Dictionary, Transformation, Reports. The architecture this chapter sits on.

→ Start here if the spine is new
— BOOK 03 · GRANT INTELLIGENCE

The Lifecycle

Same compounding architecture, applied to grant-makers. The natural companion to Book 06 for foundations that fund grantees AND operate programs.

→ For foundations & grant-makers
— BOOK 04 · IMPACT INTELLIGENCE

The Compounding

For impact funds and ESG. Nonprofits with mission-aligned investment arms or PRI portfolios live at the seam of Books 03 and 06.

→ For impact funds & ESG
— BOOK 05 · TRAINING INTELLIGENCE

The Cascade

For programs running formal training cohorts. Kirkpatrick L1–L4 with persistent learner_id. Most nonprofits running training read 04 + 06 together.

→ For training & workforce
13
§ 1.11 · SOPACT SENSE + SKILLS
— SECTION 1.11

The platform and the skill files.

Everything in this chapter runs on Sopact Sense — the data origin platform with a persistent ID at first contact. Skill files are the small Markdown recipes that turn the platform into a Theory-of-Change generator, a multi-language qualitative coder, or a partner-report aggregator. We don't distribute them; we co-author them with you in the first 60 minutes.

— THE PLATFORM

Sopact Sense

Data origin platform — not a downstream aggregator. Persistent IDs at first contact, multi-source ingest, AI analysis at submit, six reports per program cycle.

  • Multi-source intake — online · offline · docs · transcripts
  • Theory of Change generated from existing conversations
  • Persistent stakeholder_id across partners, programs, years
  • AI codes qual + quant in 40+ languages
  • Six funder-ready reports, generated continuously
Reads Airtable, KoboToolbox, Salesforce, partner PDFs. Read-only. Partners keep their stack.
— THE SKILL FILES

Co-authored at onboarding.

Four skill files cover most multi-program nonprofit work. Co-written with your team and your funder rubric — not generic templates.

toc-from-interview.md
Reads an existing transcript or strategy memo, generates a Living ToC + IRIS+ indicator map. No workshop required.
multi-language-coder.md
Codes open-ended responses, interviews, and field notes across 40+ languages. Themes cited per stakeholder_id.
partner-report-aggregator.md
Reads 100+ page partner PDFs end-to-end. Extracts metrics + themes. Surfaces cross-partner patterns at portfolio level.
board-narrative-composer.md
Synthesizes the Impact Report, Themes Report, and Variance Report into a board narrative — any output language.

Why this compounds.

A skill file is small — one or two pages of Markdown. The platform gives the skill a stakeholder_id, a Living ToC, and every wave of partner data across years. "I added two more trial prompts to the Ikaya project and I am absolutely astonished at what the system can do. And I've only just started." Marco Botha, CEO, Open Play Foundation. That's the compounding pattern — small additions, growing returns.

14
§ 1.12 · RECAP
— SECTION 1.12

Six things to take away.

Before Chapter 02 (where we build a Theory of Change from a transcript you already have), here is the compressed version of what changes when the architecture comes before the framework.

01

The frameworks are not the problem.

76% of nonprofits say measurement matters. 29% do it well. The gap is architecture, not ambition — and the field has spent fifteen years confusing the two.

02

The unread 95%.

Roughly 95% of stakeholder voice sits in PDFs, transcripts, and open-ends nobody reads. The board narrative gets built from the 5% that fits a dashboard.

03

One stakeholder_id. Many years.

A persistent ID at first contact replaces VLOOKUP, email matching, and CSV merging. The 80% cleanup tax doesn't get smaller — it disappears.

04

Start collecting, not planning.

The Living ToC emerges from conversations already happening — not a six-month workshop. Architecture first, framework second.

05

Multi-language is native.

Collect in IsiZulu. Code in English. Report in Portuguese. Same data, any language. The translation bottleneck stops being a bottleneck.

06

Continuous, not annual.

Six reports generate continuously, in any language. The funder report compresses from three-week scramble to same-day export.


— UP NEXT · CHAPTER 02

Theory of Change
From Interviews.

Chapter 02 walks through how a Living Theory of Change gets built from a transcript you already have — a partner check-in call, a coordinator field interview, a strategy memo. No workshop. No consultant. Plus: what IRIS+ alignment looks like when the indicators emerge from the data rather than precede it.

15
— END OF CHAPTER 01
— THE SOPACT INTELLIGENCE LIBRARY

Stakeholder voice is the program.

Not a footnote in the annual report. Not the open-end column that gets exported and forgotten. Stakeholder voice is the program — and the architecture should treat it that way from first contact.

THE LIBRARY · SIX VOLUMES
BOOK 01
Beyond the
Survey
Foundations
BOOK 03
Grant
Intelligence
Industry guide
BOOK 04
Impact
Intelligence
Industry guide
BOOK 05
Training
Intelligence
Industry guide
BOOK 05
Application
Management
Industry guide
BOOK 06 · NOW
Nonprofit
Programs
Industry guide

"The dashboard reads the 5%. The architecture reads the 95%. The unread is where the program actually lives."

THE SOPACT INTELLIGENCE LIBRARY · 2026
16