Breadcrumb

  1. Home
  2. Research
  3. Programs
  4. STTR: Model-Agnostic Detection of Bias (MAD-Bias)

STTR: Model-Agnostic Detection of Bias (MAD-Bias)

 

OUSD (R&E) critical technology area(s): Trusted AI and Autonomy

Objective: To develop new methods to detecting bias inherent in the datasets being used regardless of what the architecture of the downstream artificial intelligence (AI) model is.

Description: The detection of bias is conventionally framed around a dataset and an AI model that leverages said dataset for training. This topic seeks to upend that conventional wisdom and instead study the dynamics of a given dataset to detect bias, regardless of which AI model is ultimately used with that data. Potential approaches include geometric and topological data analysis, as well as alternative representations of data that facilitate an understanding of the inherent bias therein.

Phase I

The goal of Phase I proposals is to present a new technology to address AI bias as described previously. The technology need not be mature by the end of the Phase, but a convincing proof-of-concept for its utility must be demonstrated. This proof-of-concept may come in the form of a live demo, publications in peer-reviewed venues, and open-source software, among others.

Phase I deliverables and milestones for this STTR should include:

  • Month 3: report detailing technical progress made to date and tasks accomplished.
  • Month 6: finalize the technical report, including remaining challenges directions to be addressed, a tentative plan for future work, and lessons learned.

Phase II

Develop, install, integrate and demonstrate a prototype system determined to be the most feasible solution during the Phase I feasibility study. This demonstration should focus specifically on:

  1. Validating the product-market fit between the proposed solution and the proposed topic and define a clear and immediately actionable plan for running a trial with the proposed solution and the proposed customer.
  2. Evaluating the proposed solution against the objectives and measurable key results as defined in the Phase I feasibility study.
  3. Describing in detail how the solution can be scaled to be adopted widely (e.g., how can it be modified for scale).
  4. A clear transition path for the proposed solution that takes into account input from all affected stakeholders including, but not limited to: end users, engineering, sustainment, contracting, finance, legal, and cyber security.
  5. Specific details about how the solution can integrate with other current and potential future solutions.
  6. How the solution can be sustainable (i.e. supportability).
  7. Clearly identifying other specific DoD or governmental customers who want to use the solution.

Phase III dual use applications

The contractor will pursue commercialization of the various technologies developed in Phase II for transitioning expanded mission capability to a broad range of potential government and civilian users and alternate mission applications. Interested government end users may include the Air Force, the DoD Chief Digital and AI Office (CDAO), DARPA, White House Office of Science and Tech Policy (OSTP), Dept of Education, Dept of Commerce, and NIST, all of whom have been looking at the problem of detecting and mitigating bias in AI as part of an inter-agency working group. For example, mitigating bias in one of the DoD’s responsible AI principles and it is widely recognized that bias remains a hurdle for responsible AI adoption. Bias also remains a hurdle for operational AI adoption, to ensure robustness of AI to rare and unlikely events. Of course, these problems also affect and are pervasive in industry, thereby motivating the dual use of the proposed technologies. Example industrial applications include the de-biasing of generative models, which have been shown to both reflect inherent racial biases, but also to create new biases as a result of current de-biasing techniques. 

Direct access with end users and government customers will be provided with opportunities to receive Phase III awards for providing the government additional research and development, or direct procurement of products and services developed in coordination with the program. 

References

[1] Zhang, Gong, et al. "Forget-me-not: Learning to forget in text-to-image diffusion models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024. 

[2] D'Incà, Moreno, et al. "OpenBias: Open-set Bias Detection in Text-to-Image Generative Models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024.

[3] Jha, Sumit Kumar, et al. "Responsible reasoning with large language models and the impact of proper nouns" Workshop on Trustworthy and Socially Responsible Machine Learning, NeurIPS 2022. 

Keywords

AI Bias, Trustworthy AI, Trusted AI, Fair AI, Bias Mitigation

TPOC-1

DARPA BAA Help Desk

Email

SBIR_BAA@darpa.mil

Opportunity

HR0011ST2025D-05

Publication: Jan. 8, 2025
Deadline: Feb. 26, 2025

Solicitation

Resources

Contact