1328f.org

Consumer Bankruptcy Research & Accountability

← Research

Research Collaboration Brief

Methodology FJC PACER Section 1328(f) Prepared March 25, 2026

Purpose

This page consolidates the datasets, methodology, and open research questions behind the 1328f project. It is designed to pre-emptively answer the empirical questions that matter before a research conversation, not after. Every number below includes its data source and the exact query logic used to produce it.

1. The Headline Numbers

37.9M Total FJC IDB rows
(FY 2008–2024)
4.9M Chapter 13 cases
1,627,116 Prior filers flagged
(PRFILE = Y)
391,951 Prior filers discharged
(no 1328(f) check)

Source: FJC Integrated Database, public release files (civil + bankruptcy). Downloaded March 2026. Loaded into SQLite, 94 districts.

The 391,951 figure is the cross-reference of two FJC fields that had never been joined before:

Result: 27.4% of prior filers with known dispositions received a discharge. Section 1328(f) requires a time-eligibility check before discharge. No federal database records whether this check occurred.

What 391,951 is -and isn't

This is the universe of cases requiring verification, not the confirmed violation count. The FJC does not include prior case filing dates, so the aggregate data cannot determine which discharges fall within the statutory bar window (4 years for Ch. 7→13, 2 years for Ch. 13→13). Determining actual violations requires pulling individual PACER dockets and checking the dates. That is what the screener does -and why PACER access is the bottleneck.

2. Data Architecture

DatasetRecordsSourceCostAccess
FJC Integrated Database 37.9M rows fjc.gov Free Public download
Attorney outcome dataset 13,615 cases PACER API + RSS feeds ~$5,000 total Private (reproducible)
Deep-mined dockets 125 cases PACER document pulls ~$200 total Private (reproducible)
1328(f) verified violations 264 verified multi-district cases Screener + manual check ~$50 total Private (reproducible)
RECAP archive (enriched) 11,038 dockets CourtListener / Free Law Project Free Public API
RSS real-time feed ~200 new filings/day PACER RSS (free tier) Free Public

The critical asymmetry: the data showing the problem is free; the data confirming actual violations costs money. The FJC publishes the question. PACER paywalls the answer at $0.10/page.

3. Methodology: How 391,951 Was Derived

STEP 1 -Ingest

Download FJC IDB public release files (civil & bankruptcy). Two flat files covering FY 2008–2024, totaling 37.9M rows. Load into SQLite.

STEP 2 -Filter

Select Chapter 13 cases only (ORGFLCHP = '13'). Result: ~4.9M cases across 94 federal judicial districts.

STEP 3 -Identify prior filers

Filter on PRFILE = 'Y' (FJC prior-filing flag). Result: 1,627,116 cases where the debtor had at least one prior bankruptcy filing.

STEP 4 -Cross-reference dispositions

Of the prior filers, filter on disposition codes indicating discharge (D, DC, DD). Exclude cases with null or pending dispositions. Result: 391,951 prior filers received discharge.

STEP 5 -Compute rate

391,951 / 1,428,833 (prior filers with known dispositions) = 27.4% discharge rate among repeat Chapter 13 filers.

All queries run against SQLite. Schema and SQL available on request. Reproducible by anyone who downloads the same FJC files.

4. Known Limitations

These are openly acknowledged -not hidden. Several are the direct motivation for the research collaboration.

LimitationImpactMitigation
FJC lacks prior case filing dates Cannot determine from aggregate data whether discharge falls within 1328(f) bar window Screener pulls individual PACER dockets and checks actual dates. 264 verified multi-district cases verified manually.
PRFILE flag reliability unknown FJC documentation does not specify how the flag is set or its false-positive/negative rate Cross-validation against PACER case histories for the 264-case verified set shows consistent flagging.
Disposition coding inconsistency Some districts may code discharge events differently Per-district discharge rates range from 18% to 57%, suggesting real variation -but coding differences cannot be ruled out without district-level auditing.
Survivorship bias in attorney dataset The 13,615-case dataset started from one firm and expanded outward via cross-references Control groups drawn from same districts, same time periods, same chapters. Z-score: 4.5 (p < 0.00001).
PACER cost barrier limits verification scale Only 264 of 391,951 cases individually verified This is the research bottleneck. Fee-exempt institutional access would enable national verification.
Builder's case is in the dataset Potential perception of confirmation bias The 391,951 number uses zero PACER data and zero case-specific data. It is derived entirely from the government's own published database. Methodology is district-agnostic and attorney-agnostic.

5. Geographic Variation

The per-district prior-filer discharge rate varies dramatically, suggesting either enforcement variation, local practice differences, or both:

DistrictPrior Filers DischargedRateNote
D. Kansas8,21457.4%Highest in the nation
W.D. Missouri6,89144.2%Above national average
S.D. Texas14,52231.7%Near national average
N.D. Illinois22,01826.1%Near national average
N.D. Georgia11,43718.1%Lowest among large districts

Source: FJC IDB, same methodology as Section 3. Full 94-district table available.

The research question

Same statute. Same discharge bar. A 39-point spread between the highest and lowest large-district rates. Is this enforcement variation, attorney practice variation, demographic differences in repeat filing patterns, or a combination? The FJC data alone cannot answer this. PACER docket-level verification can.

6. Attorney Outcome Analysis (13,615 cases)

Independent of the FJC national analysis, a separate dataset was built from PACER API queries and RSS monitoring across 6 federal districts. This dataset tracks individual attorney outcomes over time.

13,615 Cases tracked
88.8% Dismissal rate
(subject firm)
4.5 Z-score vs. controls
p < 0.00001 Statistical significance

Control methodology

What the attorney dataset is not

This dataset is not the basis for the 391,951 finding. The two analyses are independent. The FJC analysis uses only free government data. The attorney analysis uses paid PACER data. They converge on the same structural failure -no systematic verification of discharge eligibility -from completely different starting points.

7. The Screening Tool

An open-source tool that automates Section 1328(f) eligibility verification. Given a debtor name and district, it:

  1. Queries PACER for prior bankruptcy filings
  2. Extracts filing dates and chapter types from each case
  3. Computes the statutory bar window (4 years for Ch. 7→13, 2 years for Ch. 13→13)
  4. Compares against the current case filing date
  5. Returns: eligible, barred, or insufficient data
264 Violations verified
across 7 districts
430+ Pages on
1328f.com
94 Districts
covered
#1 Google ranking
"bankruptcy discharge screener"

Live at 1328f.com. Source code on GitHub. Client-side only - no server, no database, no user data collected. Runs entirely in the browser using sql.js (SQLite compiled to WebAssembly).

The screener has been stress-tested publicly across multiple online communities, with practicing bankruptcy attorneys engaging directly on methodology. Independent users have downloaded the code and begun running it against their own districts, creating the basis for the convergence study described in Section 8.

8. Convergence Study Design (Proposed)

The most novel methodological opportunity:

The idea

A private ground-truth dataset (264 verified cases) exists. Independently, members of the public are downloading the screener from GitHub and running it against their own districts. If the crowdsourced results converge with the private dataset, the findings are independently reproducible and externally validated without any single researcher controlling the outcome.

Research design

  1. Ground truth: 264 manually verified cases (private, held by dataset builder)
  2. Independent collection: GitHub cloners running the screener against new districts
  3. Convergence test: Do independently collected results match the ground truth where districts overlap?
  4. Expansion: With institutional PACER access, scale verification from 264 verified multi-district cases to thousands -stratified by district, attorney volume, and filing year

This design separates the person who built the tool from the people validating it. If the crowd finds the same patterns independently, that is a stronger result than any single-author study.

Full convergence study design document →

9. The PACER Access Bottleneck

The core constraint

The Administrative Office of the U.S. Courts has stated that PACER fee exemptions "can only be processed for researchers affiliated with an educational institution." The dataset builder applied for an exemption and was denied on this basis. An institutional research partnership is the direct solution.

With fee-exempt access through a university affiliation:

10. Policy Impact Already Underway

DateEvent
March 17, 2026Rules suggestion submitted to uscourts.gov (Rule 4004 amendment)
March 20, 2026Formal submission to Administrative Office of the U.S. Courts
March 23, 2026Accepted as 26-BK-3 by the Advisory Committee on Bankruptcy Rules
March 23, 2026Forwarded to Advisory Committee + Standing Committee chairs and reporters
April 15, 2026Advisory Committee meeting (rules amendments on agenda)

A federal rules committee is now acting on findings that required paid PACER access to produce. The aggregate data (FJC) is free. The verification data (PACER dockets) is paywalled. This is the access-to-justice argument reduced to a concrete, active example.

The suggestion is posted on the Advisory Committee on Bankruptcy Rules suggestions page as 26-BK-3.

11. Open Research Questions

These are the questions this dataset can answer that no one has answered yet:

  1. What is the actual 1328(f) violation rate? The 391,951 is the verification universe. How many are actual violations vs. cases where the bar window had expired? Requires PACER docket-level date checking at scale.
  2. What explains the 39-point geographic spread? Is D. Kansas (57.4%) genuinely less enforcement-intensive than N.D. Georgia (18.1%), or do demographic and filing-pattern differences account for the gap?
  3. Does attorney volume predict dismissal rate? The 13,615-case dataset suggests yes (Z = 4.5). Can this hold nationally across all 94 districts?
  4. What is the dollar cost of non-verification? Each ineligible discharge that proceeds represents fees paid to an attorney, trustee payments made, and court resources consumed. Can a per-case harm estimate be built?
  5. Does the PRFILE flag have a measurable false-positive rate? Cross-validation against PACER case histories in the 264-case set suggests high accuracy, but systematic measurement has never been done.
  6. Can crowdsourced PACER verification converge with institutional verification? The convergence study design (Section 8) tests this directly.
  7. What is the optimal screening rule for Rule 4004? If the Advisory Committee adopts a verification requirement, what should the standard be? The screener provides a working prototype.

12. Reproducibility

Every finding on this page can be independently verified:

FindingHow to reproduceCost
391,951 prior filers discharged Download FJC IDB files, run the 5-step query in Section 3 Free
27.4% discharge rate Same FJC query, division step Free
Per-district rate variation Same query, grouped by district Free
264 verified violations Run screener against same case numbers on PACER ~$50
13,615-case attorney outcomes PACER API queries by attorney name + district ~$5,000
88.8% dismissal rate / Z = 4.5 Standard proportions test against same-district controls Included above

SQL schemas, Python scripts, and the full FJC query logic are available to research collaborators. The screener source code is already public on GitHub.

An invitation

This project was built by one person with no institutional affiliation, no research training, and no PACER fee exemption. Every number above is real and reproducible. What it needs now is peer review, research design expertise, and institutional access. The data is ready. The tools are ready. The rules committee is already moving.

13. Path to Nonprofit

This project is incorporating as a 501(c)(3) nonprofit. The mission: unlock the entire PACER database and make comprehensive discharge eligibility screening available to every debtor, every trustee, and every court in the United States, free of charge.

How it works

The technology works. The tools are built. The only missing piece is institutional standing. A university research partnership is the fastest path to PACER access and accelerates nonprofit formation.