Currently, the research community has been putting an effort into the development of mechanisms that improve interpretability and fairness in machine learning problems. Recent works addressed interpretability by means of model-agnostic method called SHAP. This method is based on the Shapley value, a classical concept in cooperative game theory that indicates the marginal contribution of an attribute in the output model. However, the idea of the Shapley value, as well as for other importance values, can be extended to coalitions of attributes. Therefore, we will investigate if these methods can be used to understand either the effect of single attributes or interactions between them in the trained model. Moreover, we may use these interpretations to understand how the trained model leads to unfair results.Most of the model-agnostic methods for local explanation (e.g., SHAP method) approximate the model to a linear interpretable function. In the literature, one may find capacity-based functions, such as the Choquet integral and the multilinear model, that could be used to locally fit the model and provide a clear understanding of the attributes in the output model. Indeed, the capacity coefficients in the Choquet integral are directly associated to the Shapley values. For the multilinear model, the parameters are associated to the Banzhaf values, another concept in cooperative game theory. Therefore, another goal would be to investigate if we could use these functions to improve interpretability in machine learning.It is worth mentioning that this internship research project lies in the context of the Post-Doctoral fellowship number 2020/10572-5, which comprises the investigation of novel approaches to deal with interpretability and fairness in machine learning.
News published in Agência FAPESP Newsletter about the scholarship: