TY - JOUR
T1 - Collaborative explanation of deep models with limited interaction for trade secret and privacy preservation
AU - Domingo-Ferrer, Josep
AU - Perez-Sola, Cristina
AU - Blanco-Justicia, Alberto
N1 - Publisher Copyright:
� 2019 IW3C2 (International World Wide Web Conference Committee), published under Creative Commons CC-BY-NC-ND 4.0 License.
PY - 2019/5/13
Y1 - 2019/5/13
N2 - An ever increasing number of decisions affecting our lives are made by algorithms. For this reason, algorithmic transparency is becoming a pressing need: automated decisions should be explainable and unbiased. A straightforward solution is to make the decision algorithms open-source, so that everyone can verify them and reproduce their outcome. However, in many situations, the source code or the training data of algorithms cannot be published for industrial or intellectual property reasons, as they are the result of long and costly experience (e.g. this is typically the case in banking or insurance). We present an approach whereby individual subjects on whom automated decisions are made can elicit in a collaborative and privacy-preserving manner a rule-based approximation of the model underlying the decision algorithm, based on limited interaction with the algorithm or even only on how they have been classified. Furthermore, being rule-based, the approximation thus obtained can be used to detect potential discrimination. We present empirical work to demonstrate the practicality of our ideas.
AB - An ever increasing number of decisions affecting our lives are made by algorithms. For this reason, algorithmic transparency is becoming a pressing need: automated decisions should be explainable and unbiased. A straightforward solution is to make the decision algorithms open-source, so that everyone can verify them and reproduce their outcome. However, in many situations, the source code or the training data of algorithms cannot be published for industrial or intellectual property reasons, as they are the result of long and costly experience (e.g. this is typically the case in banking or insurance). We present an approach whereby individual subjects on whom automated decisions are made can elicit in a collaborative and privacy-preserving manner a rule-based approximation of the model underlying the decision algorithm, based on limited interaction with the algorithm or even only on how they have been classified. Furthermore, being rule-based, the approximation thus obtained can be used to detect potential discrimination. We present empirical work to demonstrate the practicality of our ideas.
KW - Auditing
KW - Explainability
KW - Machine Learning
KW - Privacy
KW - Transparency
UR - https://www.scopus.com/pages/publications/85066914490
U2 - 10.1145/3308560.3317586
DO - 10.1145/3308560.3317586
M3 - Article
AN - SCOPUS:85066914490
SP - 501
EP - 507
JO - The Web Conference 2019 - Companion of the World Wide Web Conference, WWW 2019
JF - The Web Conference 2019 - Companion of the World Wide Web Conference, WWW 2019
ER -