You are here |
theprogrammersparadox.blogspot.com | ||
| | | |
donsbot.com
|
|
| | | | For the past 3 years I've been managing the Testing + Verification org at Facebook London, a group of teams that build bug finding tools, focusing on test infrastructure, dynamic analysis, automated test generation, type systems and static analysis. Now, I'm switching track to a software engineering role at Facebook, in the same organization. I'm... | |
| | | |
ssc.io
|
|
| | | | In this paper, we interrogate whether data quality issues track demographic group membership (based on sex, race and age) and whether automated data cleaning --- of the kind commonly used in production ML systems --- impacts the fairness of predictions made by these systems. To the best of our knowledge, the impact of data cleaning on fairness in downstream tasks has not been investigated in the literature. We first analyse the tuples flagged by common error detection strategies in five research datasets. We find that, while specific data quality issues, such as higher rates of missing values, are associated with membership in historically disadvantaged groups, poor data quality does not generally track demographic group membership. As a follow-up, we conduct a large-scale empirical study on the impact of automated data cleaning on fairness, involving more than 26,000 model evaluations. We observe that, while automated data cleaning is unlikely to worsen accuracy, it is more likely to worsen fairness than to improve it, especially when the cleaning techniques are not carefully chosen. Furthermore, we find that the positive or negative impact of a particular cleaning technique often depends on the choice of fairness metric and group definition (single-attribute or intersectional). We make our code and experimental results publicly available. The analysis we conducted in this paper is difficult, primarily because it requires that we think holistically about disparities in data quality, disparities in the effectiveness of data cleaning methods, and impacts of such disparities on ML model performance for different demographic groups. Such holistic analysis can and should be supported by data engineering tools, and requires substantial data engineering research. Towards this goal, we discuss open research questions, envision the development of fairness-aware data cleaning methods, and their integration into complex pipelines for ML-based decision making. | |
| | | |
simpleprogrammer.com
|
|
| | | | What Languages to Learn, How to Structure Code, Algorithms & Data Structures, Methodologies, Source Control, Object Oriented Design, Frameworks or Stack ... | |
| | | |
humanitec.com
|
|
| | A Platform Orchestrator sits at the core of an Internal Developer Platform (IDP) and enhances self-service capabilities, by giving developers the ability to deploy workloads with all dependencies to all environments, with a single workload specification. The primary function of a Platform Orchestrator is to enable Dynamic Configuration Management (DCM). |