Note
This is a Hugging Face dataset. Learn how to load datasets from the Hub in the Hugging Face integration docs.
Dataset Card for CommonForms_val#

CommonForms_val is a validation subset of the CommonForms dataset for form field detection. It contains 10,000 annotated document images with bounding boxes for three types of form fields: text inputs, choice buttons (checkboxes/radio buttons), and signature fields. This dataset is designed for training and evaluating object detection models on the task of automatically detecting fillable form fields in document images.
This is a FiftyOne dataset with 10,000 samples.
Installation#
If you haven’t already, install FiftyOne:
pip install -U fiftyone
Usage#
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = load_from_hub("Voxel51/commonforms_val_subset")
# Launch the App
session = fo.launch_app(dataset)
Dataset Details#
Dataset Description#
CommonForms_val is a validation subset extracted from the CommonForms dataset, a web-scale dataset for form field detection introduced in the paper “CommonForms: A Large, Diverse Dataset for Form Field Detection” (Barrow, 2025). The dataset frames form field detection as an object detection problem: given an image of a document page, predict the location and type of form fields.
The full CommonForms dataset was constructed by filtering Common Crawl to find PDFs with fillable elements, starting with 8 million documents and arriving at ~55,000 documents with over 450,000 pages. This validation subset contains 2,500 pages with 34,643 annotated form field instances across diverse languages and domains.
Key characteristics:
Multilingual: Approximately one-third of pages are non-English
Multi-domain: 14 classified domains, with no single domain exceeding 25% of the dataset
High-quality annotations: Automatically extracted from interactive PDF forms with fillable fields
Three form field types: Text inputs (68.9%), choice buttons (30.7%), and signature fields (0.4%)
Curated by: Joe Barrow (Independent Researcher)
Funded by: LambdaLabs (compute grant for model training)
Shared by: Joe Barrow
Language(s) (NLP): Multilingual (en, and ~33% non-English including various European and other languages)
License: [Check original repository - https://huggingface.co/datasets/jbarrow/CommonForms]
Dataset Sources [optional]#
Repository: https://github.com/jbarrow/commonforms
Paper: https://arxiv.org/abs/2509.16506
Demo: https://detect.semanticdocs.org
Original Dataset: https://huggingface.co/datasets/jbarrow/CommonForms
Uses#
Direct Use#
This dataset is intended for:
Training and evaluating object detection models for form field detection in document images
Benchmarking form field detection systems against the validation set
Research in document understanding and intelligent document processing
Developing automated form preparation tools that can convert static PDFs into fillable forms
Computer vision research on high-resolution document analysis
Multi-class object detection with imbalanced classes (signature fields are rare)
The dataset is particularly useful for:
Training YOLO, Faster R-CNN, or other object detection architectures
Fine-tuning vision transformers for document understanding
Evaluating model performance across different form field types
Studying the impact of high-resolution inputs on detection quality
Out-of-Scope Use#
This dataset should not be used for:
OCR or text recognition tasks - The dataset only contains bounding boxes for form fields, not text content
Form understanding or semantic analysis - No information about field labels, relationships, or form structure
Handwriting detection - Only detects empty form fields, not filled content
Privacy-sensitive applications without review - Forms may contain templates with sensitive field types (medical, financial, etc.)
Production deployment without validation - This is a validation subset; models should be tested on appropriate test sets
Fine-grained form field classification - Only three broad categories are available (text, choice, signature)
Dataset Structure#
FiftyOne Dataset Structure#
This dataset is stored in FiftyOne format, which provides a powerful structure for computer vision datasets:
Sample-level fields:
filepath(string): Path to the document image fileimage_id(int): Unique identifier for the image from the original datasetfile_name(string): Original filename (e.g., “0001104-0.png”)dataset_id(int): Sample ID in the original datasetground_truth(Detections): FiftyOne Detections object containing all form field annotations
Detection-level fields (within ground_truth):
label(string): Form field type - one of:text_input: Text boxes and input fields (68.9% of annotations)choice_button: Checkboxes and radio buttons (30.7% of annotations)signature: Signature fields (0.4% of annotations)
bounding_box(list): Normalized coordinates [x, y, width, height] in range [0, 1]Format: [top-left-x, top-left-y, width, height] relative to image dimensions
area(float): Area of the bounding box in absolute pixelsiscrowd(bool): COCO-style crowd flag (always False in this dataset)object_id(int): Unique identifier for the annotationcategory_id(int): Numeric category (0=text_input, 1=choice_button, 2=signature)
Image Specifications#
Image dimensions: Variable, ranging from 1680Ă—1680 to 3360Ă—3528 pixels
Mean dimensions: 1748Ă—2201 pixels
Format: RGB PNG images
Resolution: High-resolution document scans optimized for form field detection
Unique dimensions: 61 different image size combinations
Annotation Format#
Annotations follow COCO object detection format converted to FiftyOne:
Original format: COCO [x, y, width, height] in absolute pixel coordinates
FiftyOne format: Normalized [x, y, width, height] in relative coordinates [0, 1]
Bounding box validation: Invalid boxes (negative dimensions, out-of-bounds) are filtered during conversion
Dataset Creation#
Curation Rationale#
The CommonForms dataset was created to address the lack of large-scale, publicly available datasets for form field detection. Existing commercial solutions (Adobe Acrobat, Apple Preview) have limitations:
They cannot detect choice buttons (checkboxes/radio buttons)
They are closed-source and not reproducible
No public benchmarks exist for comparison
The key insight is that “quantity has a quality all its own” - by leveraging existing fillable PDF forms from Common Crawl as a training signal, high-quality form field detection can be achieved without manual annotation. This validation subset enables:
Reproducible benchmarking of form field detection systems
Open-source model development for automated form preparation
Research advancement in document understanding and intelligent document processing
Cost-effective training - models trained on this data cost less than $500 in compute
Source Data#
Data Collection and Processing#
Source: Common Crawl PDF corpus (~8 million PDFs) prepared by the PDF Association
Filtering Process:
Started with 8 million PDF documents from Common Crawl
Applied rigorous cleaning to identify well-prepared forms with fillable elements
Filtered to PDFs containing interactive form fields (text boxes, checkboxes, signature fields)
Quality filtering to ensure form fields were properly annotated in the source PDFs
Final dataset: ~55,000 documents with 450,000+ pages
Processing Steps:
PDF rendering to high-resolution images (optimized for form field detection)
Extraction of form field annotations from PDF metadata
Conversion to COCO object detection format
Train/validation/test split creation
This subset represents the validation split
Quality Assurance:
Ablation studies showed the cleaning process improves data efficiency vs. using all PDFs
Annotations are automatically extracted from interactive PDF forms (no manual annotation)
High-resolution inputs (1216px+) were found crucial for quality detection
Data Characteristics:
Multilingual: ~33% non-English pages
Multi-domain: 14 domains classified, no domain exceeds 25%
Diverse layouts: Wide variety of form designs and structures
Real-world forms: Government forms, applications, surveys, contracts, etc.
Who are the source data producers?#
The source data consists of PDF forms published on the public web and crawled by Common Crawl. The original form creators include:
Government agencies (federal, state, local)
Educational institutions
Healthcare organizations
Financial institutions
Legal services
Corporate entities
Non-profit organizations
The forms were created by professional document designers, administrative staff, and organizations worldwide. The diversity of sources contributes to the dataset’s robustness across different form styles, languages, and domains.
Note: The forms are templates (unfilled) extracted from publicly available PDFs on the internet.
Annotations#
Annotation process#
Automatic Annotation from PDF Metadata:
The annotations in this dataset are automatically extracted from interactive PDF forms, not manually annotated. The process:
Source: PDF form field metadata embedded in interactive PDFs
Extraction: Form field locations and types are programmatically extracted from PDF structure
Mapping: PDF form field types are mapped to three detection categories:
PDF text fields →
text_inputPDF checkbox/radio button fields →
choice_buttonPDF signature fields →
signature
Coordinate conversion: PDF coordinates converted to image pixel coordinates
Format standardization: Converted to COCO object detection format
Advantages:
Scale: Enables annotation of 450k+ pages without manual labor
Consistency: Annotations are objective and derived from PDF structure
Cost: No annotation costs
Quality: Reflects real-world form field placement by professional designers
Limitations:
Annotation quality depends on source PDF quality
Some PDFs may have incorrectly defined form fields
Only detects explicitly defined form fields (not visual-only fields)
Who are the annotators?#
The annotations are automatically generated from PDF metadata - there are no human annotators. The “annotators” are effectively the original form designers who created the interactive PDF forms with fillable fields.
The dataset curation and extraction pipeline was developed by Joe Barrow (Independent Researcher).
Citation#
BibTeX:
@misc{barrow2025commonforms,
title = {CommonForms: A Large, Diverse Dataset for Form Field Detection},
author = {Barrow, Joe},
year = {2025},
eprint = {2509.16506},
archivePrefix = {arXiv},
primaryClass = {cs.CV},
doi = {10.48550/arXiv.2509.16506},
url = {https://arxiv.org/abs/2509.16506}
}
APA:
Barrow, J. (2025). CommonForms: A Large, Diverse Dataset for Form Field Detection. arXiv preprint arXiv:2509.16506. https://doi.org/10.48550/arXiv.2509.16506
More Information#
Use Cases in the Wild#
The CommonForms models and dataset enable:
Automated PDF form preparation
Document digitization workflows
Accessibility improvements for forms
Form field extraction for document understanding systems