The FATE AEV track is open. Developers should submit no more frequently than
every four months, using the FATE Submission Form.
Status
[FATE, FRTE, IREX SUSPENDED]
All tracks of these evaluations will be closed from 2025-08-04 to 2025-09-08 to allow for updates to the computing infrastructure and new datasets.
[NOTE:] The duration values for algorithms submitted since 2025-09-04
are provisional. From that date NIST ran all algorithms on a more recent Linux
operating system (OS). The durations may be revised after the conclusion of
our investigation into why numeric software is slower on identical hardware
equipped with the new OS.
Facial age verification has recently been mandated in legislation in a number
of jurisdictions. These laws are typically intended to protect minors from
various harms by verifying that the individual is above a certain age, within
an age range, and, less commonly below a certain age. Other use-cases seek
only to determine actual age. While the mechanism for estimating age is usually
not specified in legislation, software-based face analysis is an attractive
approach when a photograph can be captured.
FATE AEV is an ongoing evaluation of software algorithms that inspect photos of
a face to produce an age estimate. The output is set of reports on the
accuracy and computational efficiency of algorithms. AEV is open to a worldwide
community of developers. This evaluation will remain open indefinitely as a
facility for developers to submit their algorithms whenever they are ready, but
no more frequently than four calendar months.
AE Performance
Results: [last updated: 2026-05-15]
MAE By Dataset and Variability
Mean Absolute Error (MAE) values are averages of 26 MAE estimates from two sexes and 13 ages (18 to 30).
Algorithm name
Submission data
MAE for frontal view, no-glasses, visa photos, born in Mexico.
MAE for frontal view, no-glasses, office application photos.
MAE for airport concourse border crossing photos featuring variable pose, illumination and eye glasses.
MAE for good quality mugshot portraits collected with two diffuse light sources and a uniform background.
Age Estimation Noise
Proportion of frontal video frames for age estimate was not produced
Mean absolute deviation of frontal age estimates about their mean - see
section 2.7 in the report
Proportion of left-right yaw video frames for which age estimate was not produced
Modelled increase in age estimation error for 90 degrees of head rotation - see
section 2.8 in the report
MAE Across Geography
The table shows mean absolute error (MAE) for men and women born in six regions of the
world. See the AEV report for
discussion of how these regions were selected and grouped. Lower values of MAE
indicate better accuracy. Low variation across columns indicates equitable
accuracy across countries of birth. MAE is estimated over high quality
visa-like immigration office application photos. The uncertainty estimates span
95% of bootstrap estimates of the mean. The table is ordered by the mean of
the 12 MAE ranks, but can be sorted on any column by clicking the small arrows
in its header. The columns MAX-MIN and GINI summarize demographic variability
into a single value; smaller values are better. The green shading indicates MAE
below 2 years; yellow, pink, and red colors indicate MAE values above 3.5, 4.25
and 5 years.
MAE By Interocular Distance
The results in the table below are based on measuring image resolution using
the interocular distance (IOD) instead of image width.
The values are mean absolute errors (MAE) with uncertainty estimates covering 95% of
bootstrap estimates of the mean. The photographs are mugshot photographs of white
males in the 18-30 age group collected at different venues across the United States using a
standard photographic set up but with different cameras, client-side software, and
different indviduals. The table shows MAE for six ranges of IOD measured in pixels.
Most algorithms exhibit somewhat lower MAE in images with greater interocular distances.
AEV report has
additional dataset detail. The table can be sorted on any column by clicking
the small arrows in its header.
Mean Absolute Error (MAE) by 6 ranges of IOD for white males aged 18 to 30.
[Dataset: Mugshot (high resolution)]
Age 18-30 Mean Absolute Error (MAE) values are average of thirteen MAE values for age group 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, and 30.
Age Verification
Some developers implement an age verification function that takes a photo and
legal age threshold (e.g. 18) and returns a confidence that the person in the
photo is above that age. Age verification accuracy is defined over two
populations, those underage and those overage. There is a tradeoff between two
error rates: The proportion of the underage subjects assigned a high confidence
of being overage, and the proportion of overage who are erroneously assigned a
low confidence. These depend on a confidence threshold value, which can be
adjusted by system owners.
The table shows verification accuracy for a legal age of 18. Each algorithm is
configured with a confidence threshold that allows 10% of males aged 14 to 17
to be incorrectly classified as being over 18. This means the table lists
their false positive rate (FPR) as approximately 0.1. The prior column shows
FPR for females 14-17 at that threshold. The next two columns give false
negative rates (FNR), the fraction of women and men aged 18 to 30 who are
incorrectly classified as not being above 18.
Challenge-25 & Child Online Safety (13-16)
The table shows accuracy measures for 2 common applications, Challenge-25
age-restriction and children age 13-16 online chatroom access. False positive
rates quantify how often people outside the allowed age range are estimated to
be within required category. True positive rates quantify how often children
within age 13-16 are estimated to be within that age range.
Application - False Positive Rate (FPR) values are averages of 12 FPR estimates from two sexes and six regions-of-birth for subjects aged 17.
Border - False Positive Rate (FPR) values are averages of 12 FPR estimates from two sexes and six regions-of-birth for subjects aged 17.
Child Online Safety (13-16):
[Dataset: Visa (good quality)]
Age 13-16 - Mean Absolute Errors (MAE) are averages of four MAE values for age group 13, 14, 15, and 16 (below 17).
Age 13-16 - True Positive Rates (TPR) are proportions of subjects aged 13 to 16 whose age is correctly estimated as 13 to 16 (below 17).
Age 13-16 - False Negative Rates (FNR↓) are proportions of subjects aged 13 to 16 whose age is incorrectly estimated as 8 to 12.
Age 13-16 - False Negative Rates (FNR↑) are proportions of subjects aged 13 to 16 whose age is incorrectly estimated as 17 to 22.
Age 8-12 - False Positive Rates (FPR) are proportions of subjects aged 8 to 12 but whose age is incorrectly estimated as 13 to 16 (below 17).
Age 17-22 - False Positive Rates (FPR) are proportions of subjects aged 17 to 22 but whose age is incorrectly estimated as 13 to 16 (below 17).
Resource Cost
Resource Usage:
AE Time (msec) is the median duration of the AE function call.
Config (MB) is the total size of the neural network model and configuration data of the algorithm.
Lib (MB) is the total size of the implemenation’s compiled libraries.
How to Participate
To participate in this evaluation, developers need to submit a participation
agreement to NIST, wrap their software behind the published C++ API, run their
libraries through the provided validation package (which creates a submission
package), encrypt the package, and provide a download link for the encrypted
submission package.
[Participation agreement]FATE is conducted by NIST, an agency of the United States Government. Participation is free of charge. FATE is open to a global audience of face recognition developers. All organizations who seek to participate in FATE must sign all pages of this
Participation Agreement and submit it with their algorithm submission using the Submission Form. [last update: 2023-08-17]
[API]
General and common information shared between all tracks of the FRTE/FATE evaluations are documented in a General Evaluation Specifications, which includes hardware and operating system environment, software requirements, reporting, and common data structures that support the APIs. [last update: 2023-08-17]
[Validation]
A validation package has been published. All participants must run their
software through the validation package prior to submission. The purpose of validation is to ensure consistent
algorithm output between your execution and NIST’s execution. [last update: 2023-08-17]
[Encryption]
All submissions must be properly encrypted and signed before transmission to
NIST. This must be done according to these
instructions using the
FATE Ongoing public key linked from this
page.
Participants must email their public key to NIST. The participant’s public key
must correspond to the participant’s public-key fingerprint provided on the
signed Participation Application. [last update: 2022-07-03]
[Submission]
All algorithm submissions must be submitted through the Submission Form, which requires encrypted files be provided as a download link from a generic http server (e.g., Google Drive). We cannot accept Dropbox links. NIST will not register, or establish any kind of membership, on the provided website. Participants can submit their algorithm(s), participation agreement, and GPG key at the same time via the submission form. [last update: 2023-07-03]
Participants must subscribe to the evaluation mailing list to receive emails when new reports are published or announcements are made.
Contact Information
Inquiries and comments may be submitted to [email protected].
Subscribe to the evaluation mailing list to receive emails when announcements or updates are made.