对深度学习应用的兴趣增加,以及他们的难以检测的偏见导致需要验证和解释复杂模型。然而,目前的说明方法是有限的,只要对推理过程和预测结果的解释来说都是如此。它们通常只显示模型预测很重要的图像中的位置。缺乏与解释互动的可能性使得难以确切地验证和理解模型如何工作。使用模型时,这会产生重大风险。通过解释不考虑解释的物体的语义含义,它变得复杂。为了逃避静态说明的陷阱,我们提出了一种称为Limecraft的方法,该方法允许用户交互地选择语义一致区域,并彻底检查图像实例的预测,在许多图像特征中。几种模型的实验表明,我们的方法通过检查可能表示模型偏差的图像片的模型公平来提高模型安全性。该代码可用于:http://github.com/mi2datalab/limecraft
translated by 谷歌翻译
The celebrated FedAvg algorithm of McMahan et al. (2017) is based on three components: client sampling (CS), data sampling (DS) and local training (LT). While the first two are reasonably well understood, the third component, whose role is to reduce the number of communication rounds needed to train the model, resisted all attempts at a satisfactory theoretical explanation. Malinovsky et al. (2022) identified four distinct generations of LT methods based on the quality of the provided theoretical communication complexity guarantees. Despite a lot of progress in this area, none of the existing works were able to show that it is theoretically better to employ multiple local gradient-type steps (i.e., to engage in LT) than to rely on a single local gradient-type step only in the important heterogeneous data regime. In a recent breakthrough embodied in their ProxSkip method and its theoretical analysis, Mishchenko et al. (2022) showed that LT indeed leads to provable communication acceleration for arbitrarily heterogeneous data, thus jump-starting the $5^{\rm th}$ generation of LT methods. However, while these latest generation LT methods are compatible with DS, none of them support CS. We resolve this open problem in the affirmative. In order to do so, we had to base our algorithmic development on new algorithmic and theoretical foundations.
translated by 谷歌翻译
We study the problem of explaining link predictions in the Knowledge Graph Embedding (KGE) models. We propose an example-based approach that exploits the latent space representation of nodes and edges in a knowledge graph to explain predictions. We evaluated the importance of identified triples by observing progressing degradation of model performance upon influential triples removal. Our experiments demonstrate that this approach to generate explanations outperforms baselines on KGE models for two publicly available datasets.
translated by 谷歌翻译