Optimizing Reward and Minimizing Comprehensive Cost in Recommendation Systems

Authors

  • Shuohan Gu

DOI:

https://doi.org/10.62051/msn0mj92

Keywords:

Optimizing Reward; Minimizing; Comprehensive.

Abstract

An in-depth discussion will explore the advanced methodologies designed to minimize regret in recommendation systems, shedding light on the principal strategies and assessing their effectiveness. This analysis will also consider additional factors that could profoundly impact the decision-making mechanisms within these systems. Special focus will be placed on user-specific variables such as contextual nuances, temporal dynamics, and individual preferences that, although frequently overlooked, have the potential to significantly improve the personalization and efficacy of recommendations. By incorporating these elements, the goal is to enhance the adaptability and predictive accuracy of recommendation systems, ultimately leading to a more engaging and satisfying user experience. This progressive approach is set to redefine the standards for user satisfaction and system efficiency in the dynamic realm of digital interactions and personalized technology, marking a significant evolution in how these systems cater to individual needs and preferences.

Downloads

Download data is not yet available.

References

Colas, J. T., O'Doherty, J. P., & Grafton, S. T. (2024). Active reinforcement learning versus action bias and hysteresis: Control with a mixture of experts and nonexperts. PLoS Computational Biology, 20(3), Article e1011950. https://doi.org/10.1371/journal.pcbi.1011950

Gokcesu, K., & Kozat, S. S. (2018). An online minimax optimal algorithm for adversarial multiarmed bandit problem. IEEE Transactions on Neural Networks and Learning Systems, 29(11), 5565-5580. https://doi.org/10.1109/TNNLS.2018.2806006

Yang, S., & Gao, Y. (2021). An optimal algorithm for the stochastic bandits while knowing the near-optimal mean reward. IEEE Transactions on Neural Networks and Learning Systems, 32(5), 2285-2291. https://doi.org/10.1109/TNNLS.2020.2995920

Marković, D., Stojić, H., Schwöbel, S., & Kiebel, S. J. (2021). An empirical evaluation of active inference in multi-armed bandits. Neural Networks, 144, 229-246. https://doi.org/10.1016/j.neunet.2021.08.018

Xia, Y., Qin, T., Ding, W., Li, H., Zhang, X., Yu, N., & Liu, T.-Y. (2017). Finite budget analysis of multi-armed bandit problems. Neurocomputing, 258, 13-29.

Zhu, X., Huang, Y., Wang, X., & Wang, R. (2023). Emotion recognition based on brain-like multimodal hierarchical perception. Multimedia Tools and Applications, 1-19.

Odeyomi, T. (2020). Learning the truth in social networks using multi-armed bandit. IEEE Access, 8, 137692-137701. https://doi.org/10.1109/ACCESS.2020.3012593

Auer, P., Cesa-Bianchi, N., & Schapire, R. E. (1995). Gambling in a rigged casino: The adversarial multi-armed bandit problem. Proceedings of the 36th Annual Symposium on Foundations of Computer Science, 322-331.

Elena, G., Milos, K., & Eugene, I. (2021). Survey of multiarmed bandit algorithms applied to recommendation systems. International Journal of Open Information Technologies, 9(4), 12-27.

Zhou, T., Wang, Y., Yan, L., & Tan, Y. (2023). Spoiled for choice? Personalized recommendation for healthcare decisions: A multiarmed bandit approach. Information Systems Research, 34(4), 1493-1512.

Downloads

Published

12-08-2024

How to Cite

Gu, S. (2024) “Optimizing Reward and Minimizing Comprehensive Cost in Recommendation Systems”, Transactions on Computer Science and Intelligent Systems Research, 5, pp. 855–861. doi:10.62051/msn0mj92.