Goto

Collaborating Authors

 interpretable feature selection and extraction


Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

Been Kim, Julie A. Shah, Finale Doshi-Velez

Neural Information Processing Systems

We present the Mind the Gap Model (MGM), an approach for interpretable feature extraction and selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretabil-ity and to directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation.


Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

Neural Information Processing Systems

We present the Mind the Gap Model (MGM), an approach for interpretable feature extraction and selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretability and to directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation.


Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

Neural Information Processing Systems

We present the Mind the Gap Model (MGM), an approach for interpretable feature extraction and selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretability and to directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation.


Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

Kim, Been, Shah, Julie A., Doshi-Velez, Finale

Neural Information Processing Systems

We present the Mind the Gap Model (MGM), an approach for interpretable feature extraction and selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretability and to directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation. It also maintains or improves performance when compared to related approaches. We perform a user study with domain experts to show the MGM's ability to help with dataset exploration. Papers published at the Neural Information Processing Systems Conference.


Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

Kim, Been, Shah, Julie A., Doshi-Velez, Finale

Neural Information Processing Systems

We present the Mind the Gap Model (MGM), an approach for interpretable feature extractionand selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretability andto directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation. MGM extracts distinguishing features on real-world datasets of animal features, recipes ingredients, and disease co-occurrence.It also maintains or improves performance when compared to related approaches. We perform a user study with domain experts to show the MGM's ability to help with dataset exploration.