 Final Prizes
1st Place: $25,000
2nd Place: $15,000
3rd Place: $10,000
4th place: $5,000
5th place: $3,000
4 x $1,000 awarded to the top four placements midway through the challenge (23:55 EDT, March 25, 2019). Note that teams which have passed DP prescreening will have a significant score boost.
Additional Prizes: An additional prize of $4,000 may be awarded to each of the top 5 award winning teams at the end of the final marathon match (Match #3) who agree to provide and do provide their full code solution in an open source repository for use by all interested parties, under an open source license such as BSD or Apache 2.0.
For the full breakdown of each match period, see the information provided on Challenge.gov
By registering for and competing in this challenge you consent to sharing your contact information with NIST (and agree to be contacted by NIST) in the event you are entitled to a payment on this challenge, as payments will be issued by NIST and not Topcoder.
This match is TCO19 eligible. Everybody who scores above 10k in the system test will get TCO19 points according to these rules.
Introduction
Detailed and accurate data builds the foundation for academic research, informed decision making, and a large diversity of smart software systems. Collection and sharing of relevant datasets are often complicated by various restrictions, among them the need to protect private and sensitive information contained in the dataset from 3rd parties, who still may have valid reasons to use the data. Consider a dataset listing all law enforcement incidents in a metropolitan area, where each row is an incident, and the columns contain related information such as incident location, date and time of 911 call, time of police arrival at the site, etc. There is significant benefit for independent researchers to analyze this type of public safety data for different city neighborhoods to compare police presence and public protection for different areas, etc. However, it is not acceptable to share sensitive information about specific incidents, that might allow the identification of victims, suspects, and other sensitive information about these events.
One way to secure this data is to censor all sensitive information from the dataset before making it publicly available. For example, removing exact locations of accidents, exact times, names of involved people, etc. However, this will significantly reduce the value of the remaining data for analysis. Another approach, called Differential Privacy, allows for a better balance between data privacy and the loss of useful information. The overall idea of synthetic data is the following: using the original dataset, one may generate a synthetic (also called mock, artificial) dataset that describes fictional criminal events, such that collective statistical properties of the synthetic dataset are very similar to those of the original one, and research conclusions from the synthetic dataset hold true for the original one. Differential Privacy is a formal guarantee of privacy that considers the output probability distributions of randomized, privacypreserving algorithms. Synthetic data generators that satisfy the Differential Privacy guarantee produce data that can be safely released to the public in place of the original sensitive data. We encourage you to read specialized literature and original research on differential privacy.
NIST’s (National Institute of Standards and Technology) PSCR (Public Safety Communications Research) division and Topcoder are running a series of three similar marathon matches, each twomonths long, in which you will work to create new, or improve existing, differentially private synthetic data generation algorithms. This is the third and final marathon match in the series, with a $62k prize pool. The total prize pool of the series is up to $150k which includes: $29k for the first match, $39k for the second match, plus $20k in additional prizes at the end of the final match for the winning teams who provide their full code solutions in an open source repository. In each marathon match, you will be given an original dataset, for which you will generate a synthetic dataset. We will run various statistical tests on the submitted synthetic datasets, and score them based on the proximity of their statistical properties to those of the original data.
Problem
The competitor pack provides you with Public Use Microdata Sample (PUMS) of the 1940 USA Census Data for the State of Colorado, fetched from the IPUMS USA Website. In total, ~662k records and 139 columns of data in CSV format. Check README.md document, included into the pack, for additional details on the data format.
Your goal is to develop an (ε, δ)differential privacy algorithm. You will run that algorithm on the provided dataset with an arbitrary δ ≤ 1 / (n^2), where n is the number of dataset records, and three values of ε: 8.0; 1.0; and 0.3, to produce three output CSV files. You will call these output files 8_0.csv, 1_0.csv, and 0_3.csv, pack them together into a ZIP archive, and submit the resulting ZIP to the Topcoder system for preliminary scoring, as explained below. For your convenience, the competitor pack includes Python scripts for submission generation, and a naive implementation of a simple differentially private synthetic data algorithm, as an example.
To ensure that submitted solutions were generated by valid (ε, δ)differential privacy algorithms, and with the specified values of ε and δ, you will be requested to submit a clear and complete write up of your algorithm, along with a clear and correct proof that the algorithm satisfies differential privacy, for a manual prescreening by a team of experts in differential privacy. Competitors approved during prescreening will receive ⨉1000 boost to their further provisional scores, displayed on the challenge leaderboard; i.e. submissions without such approval will have scores within the [0; 1,000] range, and submissions from prescreened competitors, having ⨉1000 factor, will score within the [0; 1,000,000] range.
The prescreening approval should not be considered as an official certification of your algorithm’s correctness of using differential privacy, although the prescreening team will do their best to ensure it. Algorithms with the potential of winning prizes will be reviewed with greater care at the conclusion of the match. To keep the leaderboard accurate, and ensure fair play during the competition, you may be requested to submit a revised writeup and pass the prescreening procedure again, to spot check the current revision of your algorithm.
Differential Privacy PreScreening Procedure
Only the team captain can submit to have your solution prescreened for differential privacy compliance (see below for teaming rulesnotably, the team captain must be a US citizen). To apply for the prescreening, the team captain should fill out this Google Form. The form requests your Topcoder username, your contact email, and a writeup containing a complete and clear algorithm specification and the theoretical proof that the algorithm satisfies differential privacy. Within approximately one week, you will be notified about the prescreening outcome. Teams that pass prescreening will be announced on the forum after each week’s prescreen review session; their leaderboard scores will receive the x1000 point boost shortly thereafter. Teams that fail prescreening will receive direct feedback by email.
Submission of Generated Privatized Datasets into Topcoder System
As explained earlier, your submission into the Topcoder system will be a ZIP archive, containing three files, named 8_0.csv, 1_0.csv, and 0_3.csv, generated with the corresponding values of ε, and δ ≤ 1 / (n^2). Each of these CSV files must be smaller than 350 Mb in size. An auxiliary Python script is included in the competitor pack, to help you with generation and check the submission archive. Once ready, you will upload this ZIP to a file sharing service that supports direct download URLs (for example, you may upload it to Google.Drive, get the sharing URL, and replace its /open segment by /uc to get the direct download URL). The resulting URL should be submitted into the Topcoder system wrapped in a simple Java code, such as this example:
class NistDp3 {
public String getAnswerUrl() {
return "https://drive.google.com/uc?id=1JV3R0fZALeqhA8DRUSscnT1GN94cGfN";
}
}
You can submit your solutions a maximum of once every two hours.
Provisional Scoring
In this match, we will use a combination of three scoring methods, explained in detail further below. The first two are the same methods used in the first match and second matches, designed to verify conservation of clustering and classification properties in the privatized datasets. The new, third method is added in this match to verify the ability of submitted solutions to conserve some derived dataset characteristics.
The overall score will be the average of the scores given by each method to each of your solutions, generated for different epsilons. Notice, that the maximal score in each method would be achieved by submission of the original dataset as the solution; of course that would violate the differential privacy, and thus won’t be a valid solution.
Scoring Method #1
A single test works on three dataset columns, picked up randomly. The domain of possible values in those columns is split into buckets. For example, say two columns are categorical with 50 and 200 possible values, and the third column is numerical, with integer values between a and b, and also permits empty values. Our scoring algorithm divides the domain of the numerical column into 100 equal ranges, of size (ab)/100 each. As the column allows empty values, it adds a “virtual range” for its empty values. Then, it creates 50 ⨉ 101 ⨉ 200 buckets, plus one special bucket for records outside of the valid value range; and for both the original and submitted datasets, it counts the number of records falling into each bucket. Resulting counts are divided by the total number of records in each dataset to get the density distribution of records. Then, our scoring algorithm calculates the absolute difference of density distributions for the original and submitted datasets, taking the sum of absolute differences of density values in corresponding pairs of buckets. Due to normalization of density distributions to 1.0, the resulting difference will be a number s, belonging to the range between 0.0 (perfect match of density distributions) and 2.0 (density distributions for the original and synthetic dataset do not overlap at all). The resulting single test score is defined as S = 1 000 000 × (1  s / 2).
To calculate the score shown in the provisional leaderboard, we created a set of 100 tests, described above, with randomly picked columns for each test. The scores from separate tests are averaged; and if the submitter has not been approved in the prescreening procedure, the score is divided by 1 000.
Scoring Method #2
A single test consist of a set of rules for different columns. Each column has 33% chance to be included into a set, thus, on average, a single test sets rules for ~11 columns. For a categorical column, the rule is a randomly picked subset of its possible values (from 1 to maximum number of values); for numeric columns, the rule is a randomly picked range of values. A dataset records satisfies the set of test rules if, in all its categorical columns that are included into the test, it has values included into corresponding rule’s subsets, and in all its numeric columns that are included into the test, it has values within the selected ranges. The way tests are generated guarantees that in the original dataset there was at least a single record matching test rules.
For the ith test we calculate the fraction of records satisfying the test rules in the original (f_{o,i}) and in the privatized (f_{p,i}) datasets. Then we quantify their mistmatch using the following formulas:
d_{i} = ln(max(f_{p,i}; 10^{6}))  ln(f_{o,i})
Δ = sqrt( (1/N) * sum_{i = 1..N}((d_{i})^2)), where N = 300 is the total number of tests
SCORE = 10^{6} max(0, 1 + Δ / (ln(10^{3})))
The score is divided by 1000 if submitter has not been approved in the prescreening.
Scoring Method #3
For each city present in the dataset (as specified by the CITY column) we calculate, based on the SEX and INCWAGE columns, Gini index and gender pay gap. We calculate the first score component based on the meansquare deviation between Gini indices obtained for the original and privatized dataset, averaged over the cities present in the original dataset. For the second score component, we rank the cities by the gender pay gap, and calculate the score component based on the meansquare deviation between the resulting city ranks in the original and privatized datasets. The overall score from the third method is the average between these two score components, and normalized to the [0; 1 000 000] or [0; 1 000] range, depending on the prescreening status of the submitter.
NOTES
 The Competitor pack includes the code for generation of such tests, and subsequent scoring of the synthetic datasets, along with instructions on how to use them for local scoring.
 It is allowed to process and submit a subset of the dataset, consisting of the first N < 34 consecutive columns. In this case, with the first scoring method, you will get a zero score for any test relying on the columns outside of the subset you have submitted. With the second scoring method, it will be assumed that any column not present in your dataset satisfies any test criteria.
Final Scoring and Validation
To determine the final ranking and prize eligibility, at the end of the provisional phase of the match, the top performers from the leaderboard will be invited (via the forum) to participate in a sequestered phase for the final scoring and validation. Invited teams will need to submit docker containers containing their executable synthetic data generators, final algorithm writeups and privacy proofs, readable source code and code guides. Note: It is very important for invited teams to watch the forum during the sequestered scoring phase. If there are problems with any of the submitted items, teams may be notified via the forum and given the opportunity to correct and resubmit. Teams will not be contacted by email to correct problems.
The final accuracy scoring will be done on 1940’s USA Census dataset for a different state, reflecting data from a different set of individuals, ensuring that the algorithm refinement process over the data made available during the provisional phase does not violate differential privacy guarantees. The first round of the final scoring will consist of testing your synthetic data generators on an independent ground truth, consisting of another set of randomly generated tests that work the same way as described above. The top performing solutions will also undergo an additional thorough differential privacy validation, including reviews of both algorithm definitions and actual source code, by the team of differential privacy experts. Final code submissions should include a short code guide (e.g. readme file) that clearly designates source code files into preprocessing, privatization and postprocessing code, and identifies where in the source code each significant algorithm step occurs. We will also run submitted solutions on the original dataset to verify that their scores were achieved with the correct ε and δ values.
We reserve the right to adjust provisional and final scoring methodologies during the contest, in a way that’s fair for all competitors, in case any flaws are found in the original protocol.
Legal Rules
This marathon match is subject to the NIST official rules. In the event of any discrepancy or inconsistency between the terms and conditions of the contest rules, the terms and conditions of the NIST Official Rules on Challenge.gov as specified within the Challenge Specific Agreement shall control. Notice the following points: teaming is allowed, participation of organizations is allowed. Participants noneligible to prizes for legal reasons, still may participate if they relinquish from the prizes.
Eligibility
To be eligible for a prize:
 Get a prize placement, with a score above 10k.
 Provide a writeup including a clear, complete and correct proof that your algorithm satisfies differential privacy.

For contestants who participated in Match #1 or Match #2 and provided solution writeups, and are continuing with the same solution strategy in Match #3, we request these contestants inform the evaluation team of differences in their Match #1 or #2 writeup by highlighting or outlining the differences.

The Official Representative of the team must be the sole person who submits the solution writeup and source code for the team, and the names in the submitted solution must clearly identify the name of the Official Rep.

Code submissions should include a short code guide (e.g. readme file) that clearly partitions their source code files into preprocessing, privatization and postprocessing code, and identifies where in the source code each significant algorithm step occurs.
 Meet the eligibility requirements in the NIST PSCR Official Challenge Rules posted on Challenge.gov.
 Per the NIST official challenge rule 3B, the official representative (individual or team lead) must be age 18 or older and a U.S. Citizen or permanent resident of the United States or its territories
 Satisfy Differential Privacy compliance review based on completion of the DP Prescreening Process (Match contest rule #8) at each match, and if you are a top scorer, submit a report for Final DP Validation (Match contest rule #9).
 Score higher than your competitors!
Teaming
Teaming is allowed in this competition along with the participation of organizations. However, there are some conditions:
 Teams can only contain 5 members including the captain.
 Team Captains must be US Citizens in order to receive prizes.
 NIST will only make one award payment per team. The payments will not be split among the members.
 The team registration form will be available in the forum after challenge launch.
References
 The Algorithmic Foundations of Differential Privacy by Cynthia Dwork and Aaron Roth.
