no code implementations • 27 Oct 2021 • Gal Yona, Daniel Greenfeld
They argue that some popular saliency methods should not be used for explainability purposes since the maps they produce are not sensitive to the underlying model that is to be explained.
no code implementations • NeurIPS 2021 • Yoav Wald, Amir Feder, Daniel Greenfeld, Uri Shalit
In this work, we draw a link between OOD performance and model calibration, arguing that calibration across multiple domains can be viewed as a special case of an invariant representation leading to better OOD generalization.
1 code implementation • ICML 2020 • Daniel Greenfeld, Uri Shalit
We adapt it to the task of learning for unsupervised covariate shift: learning on a source domain without access to any instances or labels from the unknown target domain, but with the assumption that $p(y|x)$ (the conditional probability of labels given instances) remains the same in the target domain.
1 code implementation • 25 Feb 2019 • Daniel Greenfeld, Meirav Galun, Ron Kimmel, Irad Yavneh, Ronen Basri
Constructing fast numerical solvers for partial differential equations (PDEs) is crucial for many scientific disciplines.
no code implementations • 30 Sep 2017 • Gal Hyams, Daniel Greenfeld, Dor Bank
Our suggested approaches were applied on the MNIST data-set as a proof of concept for a vision classification task and on the ADE20K data-set in order to tackle the semi-supervised semantic segmentation problem.