Sitemap

A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.

Pages

Posts

Future Blog Post

less than 1 minute read

Published:

This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.

Blog Post number 4

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 3

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 2

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 1

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

portfolio

publications

Fairness auditing through the lens of Counterfactual Explanations

Published in Under Review, 2022

A wide variety of fairness metrics and eXplainable Artificial Intelligence (XAI) approaches have been proposed in the literature to identify bias in machine learning (ML) models that are used in critical real-life contexts. In this paper, we present an approach that ties the evaluation of the fairness of machine learning models to the expense (or conversely the simplicity) of generating counterfactual explanations (CFE). Considering the relative importance of each feature class towards obtaining the desired outcome, we attribute different scores to each feature value and propose model-agnostic concepts of Qualification and Effort, respectively, indicating the candidate’s strength and the adjustments that must be made to the candidate’s profile in order to alter the classifier’s output. To ensure a fair comparison, we further group candidates based on their Qualification scores. We additionally utilize KL-divergence to capture the complex disparities between distributions of the CFE Efforts among the sensitive groups. We empirically evaluate fairness among different sensitive classes for several classifiers on two real-world datasets. We find that the majority of classifiers exhibit unfair propensities towards the minority groups. However, some models exacerbate the bias more than others. Also, we find that classifiers tend to behave differently for members of each Qualification group. Our results show that our method can successfully be used for fairness evaluation of classification models based on CFEs.

Recommended citation: Y Deldjoo, M Varasteh, YM Bahram, N Tintarev (2023). "Fairness auditing through the lens of Counterfactual Explanations" yaramohamadi.github.io.

Overfitting is not a dead-end: A survey and outlook on post-training regularization

Published in yaramohamadi.github.io (In preperation), 2023

Machine learning (ML) relies heavily on regularization, as it allows better generalization to unknown data, even with imperfect optimization procedures and datasets. There are, however, major problems with regularization that have surprisingly received little attention so far. Regularization methods traditionally avoid overfitting. But when overfitting happens, they usually fail to fight it and cannot bring the machine learning model out of its adverse situation. As a result, the stage of overfitting is taught of as a dead-end in the ML community. This idea has consequently forced the system designers to perform expensive hyperparameter searches, retraining the model from scratch every time with new configurations. It has also led to similar complications in the ever-changing dynamic usages of static pre-trained models. Defying the overfitting dead-end misconception, we argue that even a model that is overfitting includes useful information about the task at hand, and being able to adjust the regularization strength by using this information after the overfitting further solidifies this suggestion. This process ideally requires the disentanglement of the regularization process from the initial stages of the training phase and being able to apply the regularization as a post-processing step with low cost. This would allow adjusting the regularization strength of pre-trained models efficiently, which is an increasingly viable concept given the prevalent use of large neural networks today in many domains and applications. There already are implicit and explicit traces of this idea, which we call Post-Regularization, in a wide range of existing works from several domains. However, there exists no unified view of this concept. In this work, we formalize Post-Regularization, and provide a novel taxonomy of regularization, from the perspective of when the regularization is applied with respect to the model training to help bring together the ideas that can potentially be further explored in this area. We hope that this work attracts more attention to Post-Regularization and provides a foundation for future related work.

Recommended citation: YM Bahram, MA Sadeghi (2023). "Overfitting is not a dead-end: A survey and outlook on post-training regularization" yaramohamadi.github.io

talks

teaching

Personal Teaching Experience

personal, Github and Youtube, 2021

Created over 2 hours of tutorials on Machine Learning with Python in Persian language with Hands-On exercises. The tutorial covered the basics of machine learning, the primary methods and models of classification and regression, and dimensionality reduction. I published the videos on Youtube and the codes on Github.

Teaching Assistant for University Courses

university course, University of Tehran, School of Electrical and Computer Engineering, 2022

  • Deep Generative Models (Lead TA) - Fall 2022
  • Machine Learning (TA) - Fall 2022
  • Advanced Deep Learning (TA) - Spring 2022
  • Data Analytics and Visualization (Lead TA) - Fall 2021