Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Experiment grid on weights and ndkl's ratios on imdb #233

Open
Tracked by #229
Rounique opened this issue Jan 19, 2024 · 0 comments
Open
Tracked by #229

Experiment grid on weights and ndkl's ratios on imdb #233

Rounique opened this issue Jan 19, 2024 · 0 comments
Assignees
Labels
experiment Running a study or baseline for results vivaFemme

Comments

@Rounique
Copy link

Rounique commented Jan 19, 2024

*In the initial round of experiments on female_bias, an error was identified, necessitating the repetition of all experiments.

The results table encompasses various experiments employing cross-entropy, unigram, unigram_b, inverse_unigram_b, uniform, and female_bias in which diverse weights for the female_bias objective function were explored. Specifically, experiments were conducted to calculate the utility and fairness. NDKL; the fairness metric has been calculated with both males as the minority and females as the minority at various ratios. The desired distributions for female_bias, incrementing in ratios of 0.1 provide a comprehensive analysis of the impact of varying factors on the outcomes.

The results indicate a noteworthy pattern: when male experts dominate, NDKL values are superior compared to the objective functions other than female_bias. Nevertheless, as the proportion of female experts increases, NDKL exhibits improvement, even when men constitute the majority. Conversely, when females are in the majority, NDKL consistently performs significantly better across all experiments conducted with the female_bias objective function.
This implies that if the intended distribution of experts leans towards having more females, deemed as a fairer representation, utilizing objective functions other than female_bias may not effectively recommend more female experts. In contrast, employing the female_bias objective function explicitly encourages the model to prioritize and recommend more female experts, thereby aligning with the desired distribution and fostering increased female representation in the recommendation.

The experimentation involved training the model and evaluating the outcomes with varied weights for the female bias. The results unveiled a distinct trend: when men are in the majority, weights below and above 2 adversely impact NDKL. In contrast, when women are in the majority, escalating the weight correlates with an improvement in NDKL. Importantly, it is worth noting that as NDKL improves with increasing weights, there is either a minimal drop or no significant decrease in utility. This observation shows that enhancing female representation through increased weights in female bias does not compromise overall utility or may even marginally decrease it.

Results

@Rounique Rounique self-assigned this Jan 25, 2024
@hosseinfani hosseinfani added experiment Running a study or baseline for results vivaFemme labels May 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
experiment Running a study or baseline for results vivaFemme
Projects
None yet
Development

No branches or pull requests

2 participants