skip to main content
research-article

Enhancing Performance of Operationalized Machine Learning Models by Analyzing User Feedback

Published: 01 June 2022 Publication History

Abstract

Machine learning (ML) models that have been put into production must be actively monitored and maintained to ensure that the models continue to satisfy performance quality requirements. User feedback is often a very good indicator of whether the model is performing as per user expectations. We contend that monitoring ML models through user feedback, and promptly addressing the issues mentioned in the feedback will help to ensure long-term success of deployed ML models. However, the problem of incorporating user feedback for maintaining ML models has not been adequately studied. In this paper, we first motivate the importance of this problem by highlighting the benefits of incorporating user feedback. The paper discusses the challenges of effectively harnessing user feedback for enhancing performance of deployed ML models. Furthermore, we present a novel approach for analyzing user feedback to identify incompleteness of training datasets or differences between distributions of training datasets and actual workload. Our approach compares the user feedback and training data of deployed ML models to find the above mentioned deficiencies in the training dataset. And we determine a priority order to fix the issues. We also present experiments to demonstrate the effectiveness of the proposed approach.

References

[1]
Tadeusz Caliński and Jerzy Harabasz. 1974. A dendrite method for cluster analysis. Communications in Statistics-theory and Methods 3, 1 (1974), 1–27.
[2]
David L Davies and Donald W Bouldin. 1979. A cluster separation measure. IEEE transactions on pattern analysis and machine intelligence2 (1979), 224–227.
[3]
Lluís Formiga, Alberto Barrón-Cedeno, Lluis Marquez, Carlos A Henriquez, and José B Mariño. 2015. Leveraging online user feedback to improve statistical machine translation. Journal of Artificial Intelligence Research 54 (2015), 159–192.
[4]
R Stuart Geiger, Kevin Yu, Yanlai Yang, Mindy Dai, Jie Qiu, Rebekah Tang, and Jenny Huang. 2020. Garbage in, garbage out? Do machine learning application papers in social computing report where human-labeled training data comes from?. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency. 325–336.
[5]
Nitin Gupta, Shashank Mujumdar, Hima Patel, Satoshi Masuda, Naveen Panwar, Sambaran Bandyopadhyay, Sameep Mehta, Shanmukha Guttula, Shazia Afzal, Ruhi Sharma Mittal, 2021. Data Quality for Machine Learning Tasks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 4040–4041.
[6]
Ken Lang. 1995. Newsweeder: Learning to filter netnews. In Proceedings of the Twelfth International Conference on Machine Learning. 331–339.
[7]
Quoc Le and Tomas Mikolov. 2014. Distributed representations of sentences and documents. In International conference on machine learning. PMLR, 1188–1196.
[8]
Yuki Maeda, Shuji Yamaguchi, Tatsuru Higurashi, and Kota Tsubouchi. 2021. ColorfulFeedback: Enhancing Interest Prediction Performance through Multi-dimensional Labeled Feedback from Users. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining. 894–897.
[9]
Peter J Rousseeuw. 1987. Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. Journal of computational and applied mathematics 20 (1987), 53–65.
[10]
Makesh Narsimhan Sreedhar, Kun Ni, and Siva Reddy. 2020. You could have said that instead: Improving Chatbots with Natural Language Feedback. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings. 2445–2453.
[11]
Simone Stumpf, Vidya Rajaram, Lida Li, Margaret Burnett, Thomas Dietterich, Erin Sullivan, Russell Drummond, and Jonathan Herlocker. 2007. Toward harnessing user feedback for machine learning. In Proceedings of the 12th international conference on Intelligent user interfaces. 82–91.
[12]
Jinsung Yoon, Sercan Arik, and Tomas Pfister. 2020. Data valuation using reinforcement learning. In International Conference on Machine Learning. PMLR, 10842–10851.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
IVSP '22: Proceedings of the 2022 4th International Conference on Image, Video and Signal Processing
March 2022
237 pages
ISBN:9781450387415
DOI:10.1145/3531232
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 June 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. MLOps
  2. Machine Learning Operations
  3. Machine learning
  4. User Feedback for Learning

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

IVSP 2022

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 125
    Total Downloads
  • Downloads (Last 12 months)37
  • Downloads (Last 6 weeks)5
Reflects downloads up to 23 Oct 2024

Other Metrics

Citations

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media