
#8: Music Recommender Systems, Fairness and Evaluation with Christine Bauer
08/15/22 • 70 min
In episode number eight of Recsperts we discuss music recommender systems, the meaning of artist fairness and perspectives on recommender evaluation. I talk to Christine Bauer, who is an assistant professor at the University of Utrecht and co-organizer of the PERSPECTIVES workshop. Her research deals with context-aware recommender systems as well as the role of fairness in the music domain. Christine published work at many conferences like CHI, CHIIR, ICIS, and WWW.
In this episode we talk about the specifics of recommenders in the music streaming domain. In particular, we discuss the interests of different stakeholders, like users, the platform, or artists. Christine Bauer presents insights from her research on fairness with respect to the representation of artists and their interests. We talk about gender imbalance and how recommender systems could serve as a tool to counteract existing imbalances instead of reinforcing them, for example with simulations and reranking. In addition, we talk about the lack of multi-method evaluation and how open datasets incline researchers to focus too much on offline evaluation. In contrast, Christine argues for more user studies and online evaluation.
We wrap up with some final remarks on context-aware recommender systems and the potential of sensor data for improving context-aware personalization.
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from the Episode:
- Website of Christine Bauer
- Christine Bauer on LinkedIn
- Christine Bauer on Twitter
- PERSPECTIVES 2022: Perspectives on the Evaluation of Recommender Systems
- 5th FAccTRec Workshop: Responsible Recommendation
Papers:
- Ferraro et al. (2021): What is fair? Exploring the artists' perspective on the fairness of music streaming platforms
- Ferraro et. al (2021): Break the Loop: Gender Imbalance in Music Recommenders
- Jannach et al. (2020): Escaping the McNamara Fallacy: Towards More Impactful Recommender Systems Research
- Bauer et al. (2015): Designing a Music-controlled Running Application: a Sports Science and Psychological Perspective
- Dey et al. (2000): Towards a Better Understanding of Context and Context-Awareness
General Links:
- Follow me on Twitter: https://twitter.com/LivesInAnalogia
- Send me your comments, questions and suggestions to [email protected]
- Podcast Website: https://www.recsperts.com/
- (03:18) - Introducing Christine Bauer
- (09:08) - Multi-Stakeholder Interests in Music Recommender Systems
- (15:56) - Context-Aware Music Recommendations
- (21:55) - Fairness in Music RecSys
- (41:22) - Trade-Offs between Fairness and Relevance
- (48:18) - Evaluation Perspectives
- (01:02:37) - Further RecSys Challenges
In episode number eight of Recsperts we discuss music recommender systems, the meaning of artist fairness and perspectives on recommender evaluation. I talk to Christine Bauer, who is an assistant professor at the University of Utrecht and co-organizer of the PERSPECTIVES workshop. Her research deals with context-aware recommender systems as well as the role of fairness in the music domain. Christine published work at many conferences like CHI, CHIIR, ICIS, and WWW.
In this episode we talk about the specifics of recommenders in the music streaming domain. In particular, we discuss the interests of different stakeholders, like users, the platform, or artists. Christine Bauer presents insights from her research on fairness with respect to the representation of artists and their interests. We talk about gender imbalance and how recommender systems could serve as a tool to counteract existing imbalances instead of reinforcing them, for example with simulations and reranking. In addition, we talk about the lack of multi-method evaluation and how open datasets incline researchers to focus too much on offline evaluation. In contrast, Christine argues for more user studies and online evaluation.
We wrap up with some final remarks on context-aware recommender systems and the potential of sensor data for improving context-aware personalization.
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from the Episode:
- Website of Christine Bauer
- Christine Bauer on LinkedIn
- Christine Bauer on Twitter
- PERSPECTIVES 2022: Perspectives on the Evaluation of Recommender Systems
- 5th FAccTRec Workshop: Responsible Recommendation
Papers:
- Ferraro et al. (2021): What is fair? Exploring the artists' perspective on the fairness of music streaming platforms
- Ferraro et. al (2021): Break the Loop: Gender Imbalance in Music Recommenders
- Jannach et al. (2020): Escaping the McNamara Fallacy: Towards More Impactful Recommender Systems Research
- Bauer et al. (2015): Designing a Music-controlled Running Application: a Sports Science and Psychological Perspective
- Dey et al. (2000): Towards a Better Understanding of Context and Context-Awareness
General Links:
- Follow me on Twitter: https://twitter.com/LivesInAnalogia
- Send me your comments, questions and suggestions to [email protected]
- Podcast Website: https://www.recsperts.com/
- (03:18) - Introducing Christine Bauer
- (09:08) - Multi-Stakeholder Interests in Music Recommender Systems
- (15:56) - Context-Aware Music Recommendations
- (21:55) - Fairness in Music RecSys
- (41:22) - Trade-Offs between Fairness and Relevance
- (48:18) - Evaluation Perspectives
- (01:02:37) - Further RecSys Challenges
Previous Episode

#7: Behavioral Testing with RecList for Recommenders with Jacopo Tagliabue
In episode number seven, we meet Jacopo Tagliabue and discuss behavioral testing for recommender systems and experiences from ecommerce. Before Jacopo became the director of artificial intelligence at Coveo, he had founded tooso, which was later acquired by Coveo. Jacopo holds a PhD in cognitive intelligence and made many contributions to conferences like SIGIR, WWW, or RecSys. In addition, he serves as adjunct professor at NYU.
In this episode we introduce behavioral testing for recommender systems and the corresponding framework RecList that was created by Jacopo and his co-authors. Behavioral testing goes beyond pure retrieval accuracy metrics and tries to uncover unintended behavior of recommender models. RecList is an adaption of CheckList that applies behavioral testing to NLP and which was proposed by Microsoft some time ago. RecList comes with an open-source framework with ready set datasets for different recommender use-cases like similar, sequence-based and complementary item recommendations. Furthermore, it offers some sample tests to make it easier for newcomers to get started with behavioral testing. We also briefly touch on the upcoming CIKM data challenge that is going to focus on the evaluation of recommender systems.
In the end of this episode Jacopo also shares his insights from years of building and using diverse ML Ops tools and talk about what he refers to as the "post-modern stack".
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from the Episode:
- Jacopo Tagliabue on LinkedIn
- GitHub: RecList
- CIKM RecEval Analyticup 2022 (sign up!)
- GitHub: You Don't Need a Bigger Boat - end-to-end (Metaflow-based) implementation of an intent prediction (and session recommendation) flow
- Coveo SIGIR eCOM 2021 Data Challenge Dataset
- Blogposts: The Post-Modern Stack - Joining the modern data stack with the modern ML stack
- TensorFlow Recommenders
- TorchRec
- NVIDIA Merlin
- Recommenders (by Microsoft)
- recbole
Papers:
- Chia et al. (2022): Beyond NDCG: behavioral testing of recommender systems with RecList
- Ribeiro et al. (2020): Beyond Accuracy: Behavioral Testing of NLP models with CheckList
- Bianchi et al. (2020): Fantastic Embeddings and How to Align Them: Zero-Shot Inference in a Multi-Shop Scenario
General Links:
- Follow me on Twitter: https://twitter.com/LivesInAnalogia
- Send me your comments, questions and suggestions to [email protected]
- Podcast Website: https://www.recsperts.com/
Next Episode

#9: RecPack and Modularized Personalization by Froomle with Lien Michiels and Robin Verachtert
In episode number nine of Recsperts we talk with the creators of RecPack which is a new Python package for recommender systems. We discuss how Froomle provides modularized personalization for customers in the news and e-commerce sectors. I talk to Lien Michiels and Robin Verachtert who are both industrial PhD students at the University of Antwerp and who work for Froomle. We also hear about their research on filter bubbles as well as model drift along with their RecSys 2022 contributions.
In this episode we introduce RecPack as a new recommender package that is easy to use and to extend and which allows for consistent experimentation. Lien and Robin share with us how RecPack evolved, its structure as well as the problems in research and practice they intend to solve with their open source contribution.
My guests also share many insights from their work at Froomle where they focus on modularized personalization with more than 60 recommendation scenarios and how they integrate these with their customers. We touch on topics like model drift and the need for frequent retraining as well as on the tradeoffs between accuracy, cost, and timeliness in production recommender systems.
In the end we also exchange about Lien's critical reception of using the term 'filter bubble', an operationalized definition of them as well as Robin's research on model degradation and training data selection.
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from the Episode:
- Lien Michiels on LinkedIn
- Lien Michiels on Twitter
- Robin Verachtert on LinkedIn
- RecPack on GitLab
- RecPack Documentation
- FROOMLE
- PERSPECTIVES 2022: Perspectives on the Evaluation of Recommender Systems
- PERSPECTIVES 2022: Preview on "Towards a Broader Perspective in Recommender Evaluation" by Benedikt Loepp
- 5th FAccTRec Workshop: Responsible Recommendation
Papers:
- Verachtert et al. (2022): Are We Forgetting Something? Correctly Evaluate a Recommender System With an Optimal Training Window
- Leysen and Michiels et al. (2022): What Are Filter Bubbles Really? A Review of the Conceptual and Empirical Work
- Michiels and Verachtert et al. (2022): RecPack: An(other) Experimentation Toolkit for Top-N Recommendation using Implicit Feedback Data
- Dahlgren (2021): A critical review of filter bubbles and a comparison with selective exposure
General Links:
- Follow me on Twitter: https://twitter.com/LivesInAnalogia
- Send me your comments, questions and suggestions to [email protected]
- Podcast Website: https://www.recsperts.com/
- (03:23) - Introduction Lien Michiels
- (07:01) - Introduction Robin Verachtert
- (09:29) - RecPack - Python Recommender Package
- (52:31) - Modularized Personalization in News and E-commerce by Froomle
- (01:09:54) - Research on Model Drift and Filter Bubbles
- (01:18:07) - Closing Questions
If you like this episode you’ll love
Episode Comments
Generate a badge
Get a badge for your website that links back to this episode
<a href="https://goodpods.com/podcasts/recsperts-recommender-systems-experts-537441/8-music-recommender-systems-fairness-and-evaluation-with-christine-bau-69383373"> <img src="https://storage.googleapis.com/goodpods-images-bucket/badges/generic-badge-1.svg" alt="listen to #8: music recommender systems, fairness and evaluation with christine bauer on goodpods" style="width: 225px" /> </a>
Copy