Engineering Trustworthy AI : A Developer Guide for Empirical Risk Minimization

dc.contributorAalto-yliopistofi
dc.contributorAalto Universityen
dc.contributor.authorPfau, D.
dc.contributor.authorJung, Alexander
dc.contributor.departmentDepartment of Computer Scienceen
dc.contributor.groupauthorProfessorship Jung Alexanderen
dc.contributor.groupauthorComputer Science Professorsen
dc.contributor.groupauthorComputer Science - Large-scale Computing and Data Analysis (LSCA) - Research areaen
dc.contributor.groupauthorComputer Science - Artificial Intelligence and Machine Learning (AIML) - Research areaen
dc.date.accessioned2025-11-05T07:15:27Z
dc.date.available2025-11-05T07:15:27Z
dc.date.issued2025
dc.descriptionPublisher Copyright: © 2020 IEEE.
dc.description.abstractAI systems are increasingly used for critical decisions that transcend all important fields of private and public life. These systems often use empirical risk minimisation (ERM) to train powerful prediction models such as deep neural networks. The output of the predictive model runs the risk of unintentional bias, opacity, and other adverse effects. To ensure the safety of these systems, it is vital to consider these risks already in the design stage of the model. The EU acknowledged the potential sensitivity of the predictions/decisions made about persons which led to the formulation of the Ethics Guidelines for Trustworthy AI laying down seven key requirements for trustworthy AI. So far, the design of ERM-based methods prioritises accuracy over trustworthiness. This paper discusses how key requirements for trustworthy AI can be translated into design choices for ERM components. We map the design space of ML systems to the core objectives of trustworthy AI: fairness, privacy, robustness, and explainability. Our theory is instrumental in the design of trustworthy ML systems that minimise privacy leakage and are robust against (intentional) perturbations during their operation, such as disseminating fake news. The operation of trustworthy ML systems should also be transparent or explainable to its users. Finally, ML systems must be fair and not discriminate against specific user groups. There is an urgent need for a more holistic approach to ML that includes key requirements for trustworthy AI.en
dc.description.versionPeer revieweden
dc.format.mimetypeapplication/pdf
dc.identifier.citationPfau, D & Jung, A 2025, 'Engineering Trustworthy AI : A Developer Guide for Empirical Risk Minimization', IEEE Transactions on Artificial Intelligence. https://doi.org/10.1109/TAI.2025.3617936en
dc.identifier.doi10.1109/TAI.2025.3617936
dc.identifier.issn2691-4581
dc.identifier.otherPURE UUID: dda3b180-a844-4d63-9cc6-20b500713878
dc.identifier.otherPURE ITEMURL: https://research.aalto.fi/en/publications/dda3b180-a844-4d63-9cc6-20b500713878
dc.identifier.otherPURE FILEURL: https://research.aalto.fi/files/199948558/Engineering_Trustworthy_AI_-_A_Developer_Guide_for_Empirical_Risk_Minimization.pdf
dc.identifier.urihttps://aaltodoc.aalto.fi/handle/123456789/140547
dc.identifier.urnURN:NBN:fi:aalto-202511058704
dc.language.isoenen
dc.publisherIEEE
dc.relation.fundinginfoThis work was supported by Research Council of Finland grant nr. 363624, 349965 and 331197.
dc.relation.ispartofseriesIEEE Transactions on Artificial Intelligenceen
dc.rightsopenAccessen
dc.rightsCC BY
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subject.keywordAI Ethics
dc.subject.keywordEmpirical Risk minimisation
dc.subject.keywordResponsible Ai Design
dc.subject.keywordTrustworthy AI
dc.titleEngineering Trustworthy AI : A Developer Guide for Empirical Risk Minimizationen
dc.typeA1 Alkuperäisartikkeli tieteellisessä aikakauslehdessäfi
dc.type.versionpublishedVersion

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Engineering_Trustworthy_AI_-_A_Developer_Guide_for_Empirical_Risk_Minimization.pdf
Size:
680.92 KB
Format:
Adobe Portable Document Format