What this is
theAuto[Kiva] is a streaming system for abductive theory generation. A local language model reads matched pairs of Kiva microloan pitches and builds a public, evolving theory of what textual features predict faster or slower funding, and how gender interacts with those features. No human curates or edits the theory between reading and posting.
The design
Each batch draws 8 matched pairs of loans. Both loans in a pair are in the same sector with similar loan amounts, but may differ in gender and country. One funded fast (top speed quartile), one funded slow (bottom speed quartile). A pair might be two women, two men, or one of each. The model sees the gender of both loans and reasons about whether text or gender is driving the speed difference.
Women fund faster than men roughly 75% of the time on Kiva. This is the known baseline, not a discovery. The model's job is to find what in the text predicts speed beyond that baseline, and how textual features interact with gender.
What counts as a textual feature
Anything in the pitch text that a lender could read. This includes how the pitch is written (length, framing, word choice, narrative structure) and what it chooses to mention (prior repayment, children's ages, institutional affiliations, specific products, future goals). Metadata not in the pitch — loan amount, photo quality — is excluded. Sector is controlled by the matching; country and gender are visible to the model as covariates.
The proposition buckets
Propositions are filed in five buckets: faster (helps regardless of gender), slower (hurts regardless of gender), women-specific, men-specific, and interaction (works differently by gender). Gender-general propositions need evidence from multiple pair types. Interaction propositions have the highest evidence bar.
The confidence system
New propositions start at 0.55–0.65. Confidence above 0.70 is earned through multiple batches of supporting evidence. Propositions below 0.35 are auto-retired. Every 10 batches, a consolidation round fires where the model reviews its own theory. These confidence scores are heuristic triage signals, not posterior probabilities or p-values.
Methodological positioning
theAuto[Kiva] implements a streaming form of analytic induction over matched textual contrasts. The model generates testable propositions from paired cases, then revises a public rule set as new cases arrive. The closest parallels are analytic induction, the constant comparative method from grounded theory, and the logic of online rule learning. The graveyard is a built-in pruning mechanism, not a failure mode.
Surviving propositions should be interpreted as disciplined exploratory hypotheses, not formal tests of causal mechanisms. The distinctive contribution is methodological: theory revision itself becomes observable, serialized, and auditable.
Limitations
The same model both proposes and evaluates propositions. New matched pairs provide a partially adversarial evidence stream, but the tribunal of judgment is not independent. Confidence scores are uncalibrated and there is no formal error control.
Infrastructure
A collector polls the Kiva GraphQL API hourly. A runner draws matched pairs every hour. Model: gemma4:31b-it-q8_0 via Ollama. Matching, sampling, and confidence bookkeeping are computed by locked Python code that the model cannot modify. Site published via GitHub Pages.
Santosh B. Srinivas · 2026