Artificial Intelligence and Data Science in Recommendation System: Current Trends, Technologies, and Applications -  - E-Book

Artificial Intelligence and Data Science in Recommendation System: Current Trends, Technologies, and Applications E-Book

0,0
53,95 €

-100%
Sammeln Sie Punkte in unserem Gutscheinprogramm und kaufen Sie E-Books und Hörbücher mit bis zu 100% Rabatt.
Mehr erfahren.
Beschreibung

Artificial Intelligence and Data Science in Recommendation System: Current Trends, Technologies and Applications captures the state of the art in usage of artificial intelligence in different types of recommendation systems and predictive analysis. The book provides guidelines and case studies for application of artificial intelligence in recommendation from expert researchers and practitioners. A detailed analysis of the relevant theoretical and practical aspects, current trends and future directions is presented.
The book highlights many use cases for recommendation systems:
· Basic application of machine learning and deep learning in recommendation process and the evaluation metrics
· Machine learning techniques for text mining and spam email filtering considering the perspective of Industry 4.0
· Tensor factorization in different types of recommendation system
· Ranking framework and topic modeling to recommend author specialization based on content.
· Movie recommendation systems
· Point of interest recommendations
· Mobile tourism recommendation systems for visually disabled persons
· Automation of fashion retail outlets
· Human resource management (employee assessment and interview screening)

This reference is essential reading for students, faculty members, researchers and industry professionals seeking insight into the working and design of recommendation systems.

Das E-Book können Sie in Legimi-Apps oder einer beliebigen App lesen, die das folgende Format unterstützen:

EPUB

Seitenzahl: 467

Veröffentlichungsjahr: 2000

Bewertungen
0,0
0
0
0
0
0
Mehr Informationen
Mehr Informationen
Legimi prüft nicht, ob Rezensionen von Nutzern stammen, die den betreffenden Titel tatsächlich gekauft oder gelesen/gehört haben. Wir entfernen aber gefälschte Rezensionen.



Table of Contents
BENTHAM SCIENCE PUBLISHERS LTD.
End User License Agreement (for non-institutional, personal use)
Usage Rules:
Disclaimer:
Limitation of Liability:
General:
FOREWORD
PREFACE
List of Contributors
Study of Machine Learning for Recommendation Systems
Abstract
INTRODUCTION
Recommendation System
Machine Learning
Supervised learning
Semi-supervised learning
Unsupervised learning
Reinforcement learning
METHODS
Collaborative Filtering
Model-Based
Memory-Based
User-Based
Item-Based
Content-based Filtering
Hybrid Filtering
Algorithms
Co-clustering
Matrix Factorization
Singular Value Decomposition
Non-negative Matrix Factorization
Difference between SVD and NMF
K-Nearest Neighbors
K-means Clustering
Naive Bayes
Random Forest
Evaluation Methods
F1. Measure
RMSE (Root Mean Squared Error)
MAE (Mean Absolute Error)
EXPERIMENTATION
Dataset
Implementation
Result
Discussion
CONCLUSION
ACKNOWLEDGEMENT
References
Machine Learning Approaches for Text Mining and Spam E-mail Filtering: Industry 4.0 Perspective
Abstract
INTRODUCTION
Integration and Interconnection
Data and Digitalization
Refinement and Personalization
Smart Manufacturing
Automated Vehicles and Machines
Quality Control
Predictive Maintenance
Demand Predictions
Chatbots
BACKGROUND & MOTIVATION
Spam Filtering Using Machine Learning Approaches
Data Pre-processing Techniques
Spam Filtering: A Comparative Study of Machine Learning Approaches
Data Repositories
Performance Measurement
MACHINE LEARNING APPROACHES
Decision Tree Modeling
Random Forest
Gradient Boosted Model (GBM)
AdaBoost Method
Naive Bayes Classification
Artificial Neural Network
Support Vector Machines
Tuning Hyper-parameters
EXPLORATORY DATA ANALYSIS
Experimental Inferences and Discussion
CONCLUDING REMARKS
CONSENT FOR PUBLICATION
CONFLICT OF INTEREST
ACKNOWLEDGEMENT
REFERENCES
An Overview of Deep Learning-Based Recommendation Systems and Evaluation Metrics
Abstract
INTRODUCTION
RECOMMENDATION SYSTEMS
Content-based Recommendation
Collaborative Filtering Recommendation
Hybrid
DEEP LEARNING APPROACHES
Embedding
Generative Approach
Discriminative Approach
Hybrid Approach
DEEP LEARNING-BASED RECOMMENDATION SYSTEMS
Article Citation
Entertainment
E-commerce
Other Applications
EVALUATION METRICS
CONCLUSION
REFERENCES
Towards Recommender Systems Integrating Contextual Information from Multiple Domains through Tensor Factorization
Abstract
INTRODUCTION
Problem Statement
CD-CARS Overview
LITERATURE REVIEW
Cross-Domain RS
Definition of Domain
Cross-Domain Recommendation Tasks
Cross-Domain Recommendation Goals
Cross-Domain Recommendation Scenarios
Cross-domain Methods
Context-Aware Recommender Systems
Definition of Context
Obtaining Contextual Information
Contextual Information Relevance and availability
Context-Aware Approaches
“Ad-hoc” Cross-Domain Context-Aware Recommender Systems
SYSTEMATIC CROSS-DOMAIN CONTEXT-AWARE RECOMMENDER SYSTEMS
CD-CARS Problem Formalization
Contextual Information Modelling
Contextual Features Formalization
Obtaining and Choosing Relevant Contextual Information
CD-CARS Algorithms
Base Cross-Domain Algorithms
Single-Domain as CD Algorithms
Cross-Domain Algorithms
CD-CARS Evaluation
Evaluation of Data Partitioning
Sensitivity Analysis
Discussion
CONCLUSION AND RESEARCH DIRECTIONS
Acknowledgment
REFERENCES
Developing a Content-based Recommender System for Author Specialization using Topic Modelling and Ranking Framework
Abstract
INTRODUCTION
RELATED WORK
PROBLEM DESCRIPTION
HADOOP-BASED TOPIC MODELLING SYSTEM TO IDENTIFY AUTHOR SPECIALIZATION
Text Vectorization
Mapper
Reducer
INFLUENCE OF NODES AND MULTI-CRITERIA RANKING MODEL
EXPERIMENTAL SETUP AND DISCUSSION
Dataset Used
Pre-processing Step
Results of Hadoop-based Topic Modeling
Result of Ranking Model
CONCLUSION AND FUTURE SCOPE
ACKNOWLEDGEMENT
REFERENCES
Movie Recommendations
Abstract
INTRODUCTION
MOVIE RECOMMENDATION SYSTEM
RECOMMENDER SYSTEM DESIGN VARIANTS
Collaborative Filtering
Content-based Filtering
Demographic Filtering
Knowledge-based Filtering
Utility-based
Hybrid Recommender System
DESIGN OF A MOVIE RECOMMENDER SYSTEM
Machine Learning (ML) Based Approaches
Deep Learning-based Approach
THE NETFLIX RECOMMENDER SYSTEM - A CASE STUDY
Netflix Personalization
Each Row on the Page is Personalized
Ranking
PERFORMANCE METRICS ADOPTED FOR MOVIE RECOMMENDATION
CONCLUSION
REFERENCES
Sentiment Analysis for Movie Reviews
Abstract
INTRODUCTION
SENTIMENT ANALYSIS
LITERATURE SURVEY
PROPOSED WORK
Sentiment Analysis
Opinion Mining
Technical Description
Input Dataset
Dataset Description
Data Preprocessing
Deep Learning
Supervised Learning
METHODOLOGY
Random Forest
Long Short-Term Memory
Bi-Directional Long Short-Term Memory
RESULTS AND DISCUSSIONS
CONCLUSION
ACKNOWLEDGEMENT
REFERENCES
A Movie Recommender System with Collaborative and Content Filtering
Abstract
INTRODUCTION
RELATED WORK
Limitations
Proposals of a New Similarity Metrics
Accuracy
BACKGROUND
CATEGORIES OF RECOMMENDER SYSTEMS
Collaborative Recommender Systems
Memory-Based Collaborative Filtering
Model-based Collaborative Filtering
Content Recommender System
ALGORITHMS
Nearest-Neighbors
Matrix Factorization Methods
Clustering-Based RS
SIMILARITY METRICS
User-Based Collaborative Recommender System
Finding Nearest Neighbors using Jaccard Similarity
Finding Nearest Neighbors using Cosine Similarity
Nearest Neighbors using Pearson Similarity
Nearest Neighbors using Mean Square Difference Similarity
Item-Based Collaborative System
Nearest Products using Pearson Similarity
Content-Based Filters
Data Pre-processing
Vectorization
TF-IDF
Word Embeddings
Limitations
Topic Modelling
EVALUATION METRICS
Precision and Recall
MAE
RMSE
CONCLUSION AND FUTURE WORK
ACKNOWLEDGEMENTS
REFERENCES
An Introduction to Various Parameters of the Point of Interest
Abstract
INTRODUCTION
IMPACT OF VARIOUS PARAMETERS ON POI RECOMMENDATION
Users’ Interest-Based Recommendation
Location Popularity-Based Recommendation
Weather Based Recommendation
Cost Effective Recommendation
SUMMARY
CONCLUSION AND FUTURE SCOPE
Acknowledgments.
REFERENCES
Mobile Tourism Recommendation System for Visually Disabled
Abstract
INTRODUCTION
PROPOSED WORK
Recommendation Systems
Collaborative Recommender Systems
A Content-based Recommender
Hybrid Recommendation System
MAPPING TECHNOLOGIES
Tipping
Proximo
Geo Notes
Macau Map
Microsoft Planner
Tourist Guide
Cyber Guide
Context-Aware Tourist Information System
Deep Map
Tour Planning Research
Artificial Language Experimental Assistant Internet (ALEXA)
SOLUTION STRATEGY
CONCLUSION
FUTURE WORK
ACKNOWLEDGEMENT
REFERENCES
Point of Interest Recommendation via Tensor Factorization
Abstract
INTRODUCTION
Influential Factors of POI Eecommendation
Pure Check-in Based POI Recommendations
Geographical Influence Enhanced POI Recommendation
Social Influence Enhanced POI Recommendation
Temporal Influence Enhanced POI Recommendation
A Brief Introduction to Tensors
LITERATURE SURVEY ON RECOMMENDATION SYSTEM VIA TENSOR FACTORIZATION
Hotel Recommendation
Advantages
Disadvantages
Recommendation in the Travel Decision-making Process
Advantages
Disadvantages
Location-Based Social Networks for POI Recommendation
Time-Aware Preference Mining
Tensor Factorization
Advantages
Disadvantages
POI Recommendation Based on Weather Context
Context Inference and Modeling
Construction of Tensor and Feature Matrix
Time-category Matrix
Location Similarity Matrix
Location-weather Matrix
Collaborative Tensor Decomposition
POI Recommendation
Advantages
Disadvantages
POI Recommendation with Category Transition and Temporal Influence
Advantages
Disadvantages
CONCLUSION AND FUTURE SCOPE
ACKNOWLEDGMENTS
REFERENCES
Exploring the Usage of Data Science Techniques for Assessment and Prediction of Fashion Retail - A Case Study Approach
Abstract
Introduction
Previous Works
Goal and Objectives
Proposed Framework
Data Preprocessing
Feature Engineering
Predictive Analysis
Experimental Study
Data Description and Preparation
Issues and Resolution of Data
Exploratory Analysis
Feature Engineering
Impact of Rating on Sales
Impact of Material Price Season and Style on Sales
Predictive Analysis
Automation of Recommendations
Sales Forecast
Conclusion
Acknowledgement
References
Data Analytics in Human Resource Recruitment and Selection
Abstract
INTRODUCTION
RECRUITMENT ANALYTICS
Procedure for Recruitment Analytics
OPERATIONAL REPORTING
Recruiting Metrics
The Number of Days that Have Passed Since Time to Fill
Quality of Hire
Artificial Intelligence in Screening
Artificial Intelligence in Online Assessments
Artificial Intelligence in Job Interviews
Time to Hire
Cost Per Hire
First-year Attrition
Success Ratio Recruiting Metric
Employee Selection
Selection Ratio
Optimum Productivity Level (OPL)
Time to Productivity
Conclusion
Acknowledgement
REFERENCES
A Personalized Artificial Neural Network for Rice Crop Yield Prediction
Abstract
INTRODUCTION
Traditional Crop Yield Forecasting Methods
Artificial Neural Networks
LITERATURE REVIEW
STUDY AREA AND DATASET DESCRIPTION
Study Area
Dataset Description
PROPOSED METHODOLOGY
P-ANN (Personalization of ANN)
MODEL EXECUTION AND EVALUATION
Comparative Analysis
CONCLUSION AND FUTURE WORKS
ACKNOWLEDGEMENTS
REFERENCES
Artificial Intelligence and Data Science in Recommendation System:
Current Trends, Technologies and Applications
Edited By
Abhishek Majumder
Tripura University, Tripura, India
Joy Lal Sarkar
Tripura University, Tripura, India
&
Arindam Majumder
NIT Agartala, Tripura 799046, India

BENTHAM SCIENCE PUBLISHERS LTD.

End User License Agreement (for non-institutional, personal use)

This is an agreement between you and Bentham Science Publishers Ltd. Please read this License Agreement carefully before using the ebook/echapter/ejournal (“Work”). Your use of the Work constitutes your agreement to the terms and conditions set forth in this License Agreement. If you do not agree to these terms and conditions then you should not use the Work.

Bentham Science Publishers agrees to grant you a non-exclusive, non-transferable limited license to use the Work subject to and in accordance with the following terms and conditions. This License Agreement is for non-library, personal use only. For a library / institutional / multi user license in respect of the Work, please contact: [email protected].

Usage Rules:

All rights reserved: The Work is the subject of copyright and Bentham Science Publishers either owns the Work (and the copyright in it) or is licensed to distribute the Work. You shall not copy, reproduce, modify, remove, delete, augment, add to, publish, transmit, sell, resell, create derivative works from, or in any way exploit the Work or make the Work available for others to do any of the same, in any form or by any means, in whole or in part, in each case without the prior written permission of Bentham Science Publishers, unless stated otherwise in this License Agreement.You may download a copy of the Work on one occasion to one personal computer (including tablet, laptop, desktop, or other such devices). You may make one back-up copy of the Work to avoid losing it.The unauthorised use or distribution of copyrighted or other proprietary content is illegal and could subject you to liability for substantial money damages. You will be liable for any damage resulting from your misuse of the Work or any violation of this License Agreement, including any infringement by you of copyrights or proprietary rights.

Disclaimer:

Bentham Science Publishers does not guarantee that the information in the Work is error-free, or warrant that it will meet your requirements or that access to the Work will be uninterrupted or error-free. The Work is provided "as is" without warranty of any kind, either express or implied or statutory, including, without limitation, implied warranties of merchantability and fitness for a particular purpose. The entire risk as to the results and performance of the Work is assumed by you. No responsibility is assumed by Bentham Science Publishers, its staff, editors and/or authors for any injury and/or damage to persons or property as a matter of products liability, negligence or otherwise, or from any use or operation of any methods, products instruction, advertisements or ideas contained in the Work.

Limitation of Liability:

In no event will Bentham Science Publishers, its staff, editors and/or authors, be liable for any damages, including, without limitation, special, incidental and/or consequential damages and/or damages for lost data and/or profits arising out of (whether directly or indirectly) the use or inability to use the Work. The entire liability of Bentham Science Publishers shall be limited to the amount actually paid by you for the Work.

General:

Any dispute or claim arising out of or in connection with this License Agreement or the Work (including non-contractual disputes or claims) will be governed by and construed in accordance with the laws of Singapore. Each party agrees that the courts of the state of Singapore shall have exclusive jurisdiction to settle any dispute or claim arising out of or in connection with this License Agreement or the Work (including non-contractual disputes or claims).Your rights under this License Agreement will automatically terminate without notice and without the need for a court order if at any point you breach any terms of this License Agreement. In no event will any delay or failure by Bentham Science Publishers in enforcing your compliance with this License Agreement constitute a waiver of any of its rights.You acknowledge that you have read this License Agreement, and agree to be bound by its terms and conditions. To the extent that any other terms and conditions presented on any website of Bentham Science Publishers conflict with, or are inconsistent with, the terms and conditions set out in this License Agreement, you acknowledge that the terms and conditions set out in this License Agreement shall prevail.

Bentham Science Publishers Pte. Ltd. 80 Robinson Road #02-00 Singapore 068898 Singapore Email: [email protected]

FOREWORD

I have the pleasant task of writing the foreword for the book Artificial Intelligence and Data Science in Recommendation System: Current Trends, Technologies, and Applications. This is a work edited by Abhishek Majumder, Joy Lal Sarkar of Tripura University, India, and Arindam Majumder of NIT Agartala, India. This book spans certain very crucial and current issues on the theory and application of Artificial Intelligence and Machine Learning. One of the most widely used applications is recommendation systems, which millions of people use on an everyday basis for shopping and entertainment.

The methods in AI and NLP have been in development for several decades. Classification methods and neural networks have also existed for a long time. However, the advent of large-scale gathering of social and user data has recently allowed theoretical techniques to be tested and proved in everyday practice. As a student of AI in the late 80s at IISc, it was difficult for me to imagine this day. We have seen the progression of the methods of pattern recognition and statistical classification methods. There was an interesting twist in the developments of AI systems where in the late 60s, it appeared that linear classification systems and Perceptron training algorithms would progress far. But the failure to solve XOR logic problems led the researchers to believe that these would be ineffective. This has now been very much established to be a fallacy. But the twist took AI research into the development of logic and systems called expert systems. It was imagined that these expert systems would have the real world and the real world experts' knowledge. The knowledge acquisition bottleneck and lack of trainability of the expert systems were their downfalls. There is now a resurgence of another type of system that is filling in this role: the recommender system. These systems are bringing together diverse methods and techniques in AI, Data Science, and large data sets into human interfaces.

Thus, it gives me immense pleasure to see that this compilation has various applications, such as industry 4.0. Going further, we have applications presented here on deep learning, developing applications, movie recommendations, and movie reviews. One of the major applications these days is through natural language processing methods to perform sentiment analysis with data from social media. This is applied to movie reviews for tourist reviews, assessment of prediction and fashion retail, and exploring human resource recruitment and selection aspects. In addition to the very current topics that have been compiled, it is seen that there is a good diversity in the contributors to this volume.

I wish this compilation the best wishes and that the readers might benefit most from it.

Atul Negi Professor School of Computer and Information Sciences University of Hyderabad Hyderabad India

PREFACE

A recommendation System is an intelligent computer-based system that serves as a guide and suggests, as per the preferences of the person. It uses state-of-the-art technologies like Big Data, Machine Learning, Artificial Intelligence, etc., and benefits both the consumer and the merchant. Recommendation System is becoming very popular as it serves as a guide for the activity that a person or a group plans to perform in the best possible manner, given the constraints imposed by the user(s). Software tools and techniques provide advice on items to be used by a user. The recommendations are to inspire its users to buy different products. This music creation initiative includes specialists in several fields, including Artificial Intelligence, Human-Computer Interaction, Data Mining, Analytics, Adaptive User Interfaces, and Decision Support Systems, etc. In this book, the major concepts of recommender systems, theories, methodologies, challenges and advanced applications of recommenders systems are imposed on this diversity. This book comprises various parts: techniques, applications and assessments of recommendation systems, interactions with these systems, and advanced algorithms. The topic of recommendation systems is highly diverse, since it makes it possible for users to make recommendations using different types of user preferences and user needs data. Collaborative filtering processes, content-based methods, and knowledge-based methods are the most common methods in recommending systems. Such three approaches are the basic foundations of recommendation systems. Specialized methods for different data fields and contexts, such as time, place, and social information, have been developed in recent years. Many developments for specific scenarios have been suggested, and techniques have been adapted to different fields of use.

Abhishek Majumder Tripura University Tripura IndiaJoy Lal Sarkar Tripura University Tripura India &Arindam Majumder National Institute of Technology Agartala Tripura 799046

List of Contributors

Abdul WahidAmity Global Institute, Singapore 238466, SingaporeAbhishek MishraIndian Institute of Technology, Bhubaneswar, IndiaAbhishek MajumderMobile Computing Lab, Department of Computer Science and Engineering, Tripura University, Tripura, IndiaAnukampa BeheraDepartment of Computer Science & Engineering, ITER, S'O'A (Deemed to be) University, Bhubaneswar, IndiaAlladi SureshbabuDepartment of CSE, JNTUA College of Engineering, Ananthapur, Andhra Pradesh, IndiaAnupama AngadiDepartment of Information Technology, Anil Neerukonda Institute of Technology & Science, Visakhapatnam, Andhra Pradesh, IndiaAndré NascimentoDepartment of Computing, Federal Rural University of Pernambuco, Recife, BrazilBalajee MaramDepartment of Computer Science and Engineering, GMR Institute of Technology (Autonomous), Rajam, Andhra Pradesh, IndiaBibudhendu PatiDepartment of Computer Science, Rama Devi Women's University, Odisha, IndiaDillip RoutDepartment of Computer Science and Engineering, Centurion University of Technology and Management, Odisha, IndiaChhabi Rani PanigrahiDepartment of Computer Science, Rama Devi Women's University, Odisha, IndiaDouglas VérasDepartment of Computing, Federal Rural University of Pernambuco, Recife, BrazilGustavo CallouDepartment of Computing, Federal Rural University of Pernambuco, Recife, BrazilGoddumarri Surya NarayanaDepartment of CSE, Vardhaman College of Engineering, Hyderabad, TS, IndiaJoy Lal SarkarMobile Computing Lab, Department of Computer Science and Engineering, Tripura University, Tripura, IndiaKhushi ChavanDepartment of Computer Engineering, Dwarkadas J. Sanghvi college of Engineering, Mumbai, Maharastra, IndiaKotrike Rathnaiah RadhikaDepartment of Information Science and Engineering, BMS College of Engineering, Bangalore, IndiaPadmaja PoosapatiDepartment of Information Technology, Anil Neerukonda Institute of Technology & Science, Visakhapatnam, Andhra Pradesh, IndiaPundru Chandra Shaker ReddyDepartment of CSE, CMR College of Engineering & Technology, Hyderabad, TS, IndiaPooja SelvarajanDepartment of Computer Science and Engineering, Sona College of Technology, IndiaPoovizhi SelvanDepartment of Computer Science and Engineering, Sona College of Technology, IndiaPradeep KumarDepartment of CS&IT, Maulana Azad National Urdu University, Hyderabad, IndiaRamchandra MangrulkarDepartment of Computer Engineering, Dwarkadas J. Sanghvi college of Engineering, Mumbai, Maharastra, IndiaRajesh BhatiaDepartment of Computer Science & Engineering, Punjab Engineering College, Chandigarh, IndiaSatya Keerthi GorripatiComputer Science and Engineering, Gayatri Vidya Parishad College of Engineering (Autonomous), Visakhapatnam, Andhra Pradesh, IndiaSumi Kizhakke ValiyatraInstitute of Management in Kerala, University of Kerala, Kerala 695034, IndiaSuneetha MerugulaDepartment of Information Technology, GMR Institute of Technology, Rajam, Andhra Pradesh, IndiaSandeep HaritDepartment of Computer Science & Engineering, Punjab Engineering College, Chandigarh, IndiaSathiyabhama BalasubramaniamDepartment of Computer Science and Engineering, Sona College of Technology, Tamilnadu, IndiaSanthosh Kumar BalanDepartment of Computer Science & Engineering, Guru Nanak Institute of Technology, Hyderabad, TelanganaShilpa VermaDepartment of Computer Science & Engineering, Punjab Engineering College, Chandigarh, IndiaShreya RoyMobile Computing Lab, Department of Computer Science and Engineering, Tripura University, Tripura, IndiaSumit MitraManaging Partner, Citizen, Odisha, IndiaSamudrala Venkatesiah SheelaDepartment of Information Science and Engineering, BMS College of Engineering, BangaloreTushar DeshpandeDepartment of Computer Engineering, Dwarkadas J. Sanghvi college of Engineering, Mumbai, Maharastra, IndiaVidhushavarshini SureshkumarDepartment of Computer Science and Engineering, Sona College of Technology, IndiaVenkatesh NaganathanAmity Global Institute, Singapore, SingaporeYadala SucharithaDepartment of CSE, CMR Institute of Technology, Hyderabad, TS, India

Study of Machine Learning for Recommendation Systems

Tushar Deshpande1,*,Khushi Chavan1,Ramchandra Mangrulkar1
1 Department of Computer Engineering, Dwarkadas J. Sanghvi college of Engineering, Mumbai, Maharastra, India

Abstract

This study provides an overview of recommendation systems and machine learning and their types. It briefly outlines the types of machine learning, such as supervised, unsupervised, semi-supervised learning and reinforcement. It explores how to implement recommendation systems using three types of filtering techniques: collaborative filtering, content-based filtering, and hybrid filtering. The machine learning techniques explained are clustering, co-clustering, and matrix factorization methods, such as Single value decomposition (SVD) and Non-negative matrix factorization (NMF). It also discusses K-nearest neighbors (KNN), K-means clustering, Naive Bayes and Random Forest algorithms. The evaluation of these algorithms is performed on the basis of three metric parameters: F1 measurement, Root mean squared error (RMSE) and Mean absolute error (MAE). For the experimentation, this study uses the BookCrossing dataset and compares analysis based on metric parameters. Finally, it also graphically depicts the metric parameters and shows the best and the worst techniques to incorporate into the recommendation system. This study will assist researchers in understanding the summary of machine learning in recommendation systems.

Keywords: F1-measure, Machine learning, Mean absolute error (MAE), Nearest k- neighbors (KNN), Non-negative matrix factorization (NMF), Recommendation system, Root mean squared error (RMSE), Singular value decomposition (SVD).
*Corresponding author Tushar Deshpande: Department of Computer Engineering, Dwarkadas J. Sanghvi college of Engineering, Mumbai, Maharastra, India; Tel: +91-07599029823; E-mail: [email protected]

INTRODUCTION

Recommendation System

The recommendation system [1] is the main part of digitization as it analyses the interest of users and recommends something based on those interests [2-5]. The aim of these systems is to reduce information overload by retrieving the most sim-

ilar items depending on the customer's interest [6-10]. The primary use of these systems is decision making, maximizing profits, and reducing risks. This reduces customer’s efforts and time in information searching. It works as a filter that suggests alternatives based on massive data. Moreover, it acts as a multiplier that contributes to the expansion of the client’s options [11-22].

Over the last few years, the enthusiasm for recommendation systems has increased tremendously [23]. This is the most widely used service on high-end websites like Amazon, Google, YouTube, Netflix, IMDb, TripAdvisor, Kindle, etc. A number of media companies develop these systems as a service model for their clients. Furthermore, the implementation of such systems at commercial and non-profit sites attracts the attention of the customer [24-32]. These also satisfy clients more with online research results. These systems help customers search for their loved items faster and acquire more authentic predictions leading to higher sales at an eCommerce site.

Regarding knowledge of these systems, there are various undergraduate and graduate courses at institutions around the world. Conferences, workshops, and contests are organized in accordance with these systems [33-47]. One of the competitions was the Netflix Prize, organized around machine learning and data mining. In this competition, participants were required to develop a movie recommendation system whose accuracy is 10% more precise than the existing system, also known as Cinematch. After a year of hard work, the Korbell team won first place using the two main algorithms: matrix factorization (Singular value decomposition (SVD)) and Restricted Boltzmann machines (RBM).

Real applications [2] employ different ML algorithms, such as K-nearest neighbor (KNN), Naive Bayes, Random Forest, Adaboost, Singular value decomposition (SVD), and many others. The evolution of the recommendation scheme has led to the application of ML and AI algorithms for effective prediction and accuracy. In addition, the results provided by some ML algorithms are expected to be slightly promising. Due to the broad classification of ML algorithms, the choice of an ML algorithm may become a challenge depending on the different situations where recommendation systems are needed. To select an effective ML algorithm, the best way for the researcher or programmer would be to have a thorough knowledge of ML and recommending systems [48, 49]. This knowledge enables the researcher to create a model appropriate to a specific problem. Here, the study provides an overview of ML briefly.

Machine Learning

Machine learning demonstrates the imitation of human learning in computers by learning from experiences and applying them to recently encountered situations. ML originated in the 1950s but became more popular in the 1990s. Humans understand, but on the other side, the computer uses algorithms.

Machine Learning is classified into four categories:

1. Supervised learning

2. Semi-supervised learning

3. Unsupervised learning

4. Reinforcement learning

Supervised learning

This learning deals with algorithms that provide training data with a set of features and the correct prediction according to those features. The task of the model would be to learn from this data and apply the information learned into new data with the input features and predict its outcome. An example would be predicting the price of a house according to the area.

Semi-supervised learning

In this learning, the model learns from training data that includes missing information. These types of algorithms focus more on concluding from insufficient data. An example is the evaluation of movies where not all viewers will give a review, but the model ends with the reviews provided.

Unsupervised learning

This learning focuses on algorithms that do not require training data. These algorithms use real-world information to learn by themselves. It focuses primarily on relations hidden in the specified data. An example is YouTube, which parses the viewed videos and recommends similar videos to the user.

Reinforcement learning

This type of learning involves algorithms that learn from feedback from an external body. It is similar to a student and teacher where the teacher may give fewer grades (negative feedback) or more grades (positive feedback). An example is to offer a treat to a dog for a positive response and not give that treat for a negative one.

METHODS

The idea of recommendation systems is to provide recommendations to the user according to their behavior or profile. It analyzes the user's interest dynamically so that when the user carries out actions, he recommends according to his tastes. Various types of recommendations also involve recommendations based on trust, context, and risk. The types discussed in this document can be found in Fig. (6). The Recommendation System [4] is mainly divided into three categories:

1. Collaborative filtering

2. Content-based filtering

3. Hybrid filtering

Collaborative Filtering

In this approach [5], recommendation systems work according to user information. It compares users of similar preferences and recommends trying items that other users have tried shown in Fig. (1). An example is book applications in which the model would search for similar preferred users and would recommend what was purchased by those users to the current user. This type of system is further divided into a memory-based and model-based approach The Difference between memory-based and model-based method is shown in Fig. (2).

Fig. (1)) Example of Collaborative filtering [6].

Model-Based

In this method [7], the information base is past evaluations by which the model learns for better future predictions. This method functions on items that are not yet seen or used by the user. This method increases the accuracy of the system. Model-based approaches include matrix factorization, clustering, association techniques, Bayesian networks, and many more.

Memory-Based

In this method, the basis of the information is the likes and dislikes of other users, which is similar to the profile of the user who requires recommendations. This approach analyses the similarity between user interests to predict an item to the desired user. The approach is divided into subtypes, particularly user-based and item-based methods Fig. (3) shows the difference between user-based and item-based method. .

User-Based

This approach analyses the similarity among users in predictions. It can also predict, depending on the desired user's behavioral patterns. For example, if a user purchases a book, they will analyze other users' preferences on that book and recommend new items to the user.

Item-Based

This approach analyzes the similarity between the items researched or purchased by users for predictions. In other words, it computes the similarities between items unknown to the user and items known to the user and displays unknown items if the similarity value is high. For example, if a user buys an item, this system will look for items with similar features to the item purchased and recommend it to the user.

Fig. (2)) Difference between memory-based [8] and model-based [9]. Fig. (3)) Difference between user-based and item-based [10].

Content-based Filtering

In this approach, the recommendation system functions based on the data of the item the user is looking for. The model analyses other items with attributes similar to those in the search and recommends them to the user. An example, shown in Fig. (4), is online shopping, where the user searches for an item with specific features and recommends similar items.

Fig. (4)) Example of Content-based filtering [6].

Hybrid Filtering

This approach is a combination of the two earlier methods, as illustrated in Fig. (5). This means that these recommendation systems are based on item data and user information. The first step consists of analyzing the user information. The second step is to analyze the data element you are looking for or using. Finally, the relevant dataset of the first two steps appears in the form of recommendations (Fig. 6).

Fig. (5)) Mechanism of Hybrid filtering. Fig. (6)) Tree diagram of Filtering Techniques.

Algorithms

This article includes a detailed explanation of Singular value decomposition (SVD), Non-negative matrix factorization (NMF), K-means clustering, K-nearest neighbors (KNN), Co-clustering, Naive Bayes, and Random Forest algorithms.

Co-clustering

Co-clustering, also known as bi-clustering [11], is a method wherein there is a simultaneous clustering between rows and columns of a matrix. This matrix represents information as a function of user characteristics and item characteristics. In other words, co-clustering can also be visualized as grouping two different kinds of entities according to their similarity. The result of a co-clustering algorithm is commonly termed a bi-cluster [12, 13]. The kinds of bi-clustering are classified according to the nature of these bi-clusters. It depends mainly upon constant and consistent values.

1) Bi-cluster with constant values: Rows and columns within a clustering block have the same constant value.

2) Bi-cluster with constant values in rows or columns: Every row or column in a clustering block has the same constant value.

3) Bi-cluster with coherent values: These bi-clusters identify more complex similarities between genes and conditions using an additive or multiplicative method.

It is used across a wide variety of applications. Rege et al. [14] use co-clustering for clustering documents and topics. Chen et al. [15] and Felzenszwalb and Huttenlocher [16] use image co-clustering for image processing. It also helps to identify interaction networks [17, 18]. It is also an analytical tool for election data. The clustering technique is implemented through a variety of matrix factorization techniques.

Matrix Factorization

Matrix factorization is a type of algorithm associated with the decomposition of the user-item interaction matrix into the product of two rectangular matrices. This is usually done by minimizing the mathematical cost function RMSE (Root mean square error) which is done using gradient descent. Because of its effectiveness, this method became more popular during the Netflix Prize challenge (as discussed above). Recommendation systems use different matrix factorization techniques. Furthermore, a detailed study on Singular value decomposition (SVD) and Non-negative matrix factorization (NMF) is given below.

Singular Value Decomposition

This method is associated with linear algebra and is increasingly popular within ML algorithms. Its application is mainly recommendation systems for e-commerce, music, or video streaming sites.

SVD refers to the decomposition of a single matrix into three additional matrices. The general form is:

(1)

where M is the given mxn matrix,

X is an mxn orthogonal matrix that denotes the relation between the user and latent factors,

S is an nxn diagonal matrix that denotes the strength of these latent factors, and

Y is nxn orthogonal matrix and it represents the similarity between the user and latent factors.

The steps involved in SVD are given below:

1. In the first step, the data is represented as a matrix with rows as user and columns as items.

2. If there are any empty entries in the matrix, provide the average of the other entries so that there is no major error in the calculation.

3. After this, compute the SVD. (Done using numpy and surprise library)

4. After calculating the SVD, you only need to reduce it to obtain the expected matrix that will be used for the prediction by looking at the appropriate user/article pair.

The primary benefit of SVD is that it simplifies the data set and eliminates noise from the data set. It also functions with the numerical data set. Also, it could improve the precision. There are many issues related to the SVD. One of the most important issues is data scarcity, also called the cold start problem [20]. This occurs due to a new community, user, or item. If a new community, user, or item is added, the recommendation system will not work properly due to a lack of information. Black sheep is also an issue, meaning some customers also agree and disagree with the same group of people. If so, it is impossible to make recommendations. Due to its temporal complexity (O (n)), it also suffers from scalability issues.

There are different applications of SVD. The most common applications are pseudo- inverse, resolving homogeneous linear equations, minimizing total least squares, range, null space and rank, and approximation to the lowest rank matrix. In addition, it is used for signal processing, image processing, and big data.

Non-negative Matrix Factorization

This is also a matrix factorization technique [21]. As with SVD, the analogy for this approach is to break down or factorize a given matrix. The only difference, on the other hand, is that the matrix is split into two parts. The two parts are called W and H. W matrix is for weights which represent each column as a basic element. These are building blocks from which to obtain predictions to the original data item. H matrix is hidden, which represents the coordinates of the data items of W. In other words, it guides us in converting to the original data item from the group of building blocks of W.

The order of execution in NMF is given below:

1. Import the NMF model using the surprise library.

2. Then, load the dataset and isolate it to the given model.

3. Later, clean the data and create a function to pre-process data.

4. Successively create a document term matrix 'V'(given matrix).

5. Create a function to display the mode features.

6. Then, run NMF on the document term matrix 'V'.

7. Continue checking and iterating until useful features are found.

The advantage of NMF is that it breaks down the given matrix into two smaller matrices whose dimensions can be controlled by the given matrix. It differs from other matrix factorization algorithms because it works only on positive numbers which makes the data interpretable. The dataset can become smaller if W and H are depicted sparsely. The issue with the semi-supervised NMF is that depending on the number of data points available, there is a reduction in the fitted data points.

Applications of the NMF include the processing of audio spectrograms, document clustering, recommendation systems, chemometrics, and many others. It is also used for dimensionality reduction in astronomy, statistical data imputation, as well as nuclear imaging.

Difference between SVD and NMF

So as stated above, both SVD and NMF are matrix factorization techniques. But there are also some differences between them, which could help us to choose the best algorithm for a situation between these two.

1. The SVD includes both negative and positive values, while the NMF has strictly positive values. That makes NMF useful because it provides more sense and connections are made easier.

2. SVD factors can be related to the eigenfunctions of a system where the original matrix denotes a system about which one is taking interest from a signal processing perspective. This makes SVD more effortless. Although NMF can also be used for the same purpose because the association is indirect in this approach, it becomes more tedious.

3. The factors of SVD are unique, whereas the factors of NMF are not unique. As a result, NMF is better for algorithms with privacy protection.

4. SVD factors into three matrices, out of which the sigma matrix gives the information stored in the vector. Whereas NMF only factors into two matrices which do not include the sigma matrix.

K-Nearest Neighbors

KNN is an easy machine learning algorithm based on supervised ML learning. It finds similar items based on the distance between test data and individual training data using a variety of distance concepts. In this algorithm, predictions are mainly made using the calculation of the Euclidean distance of the nearest neighbors. Besides, the use of Jaccard similarity, Minkowski, Manhattan, or Hamming distance can be done instead of Euclidean. This is a non-parametric algorithm that assumes nothing about the given data. It is also referred to as a lazy learning algorithm, which does not learn from data, but instead stores and performs actions on the data.

The steps involved in KNN are given below:

1. Load the dataset and preprocess it.

2. Fit the KNN algorithm (defined as Nearest-Neighbors) to the training dataset (use the sklearn library). For using the surprise library, it is defined as KNNBasic.

3. Predict the test result.

4. Creating the confusion matrix and finding the test accuracy of the result.

5. After this, the visualization of the test result can be done.

This algorithm is used as it is easy to interpret the result. It also has great predictive power and less computing time. The main issue with KNN is that it becomes much slower as the volume of data increases. As such, it does not give good accuracy with large datasets. It is also highly sensitive to missing values, outliers, and noise from the dataset.

It is primarily used for classification and regression problems. The result of a classification problem is a discrete value while for a regression problem, the result is a real number (containing a decimal). It is commonly used for text extraction. It is used in finance for stock prediction, management of loans, and analysis of money laundering. It is used in agriculture for weather forecasting and estimation of soil water parameters. It is also used in medicine to predict different diseases.

K-means Clustering

The k-means algorithm is the most widely known clustering algorithm. It is the simplest method of unsupervised learning to resolve the clustering issue. It also aims at solving the Expectation-Maximization problem. In this algorithm, a k value is received that represents the number of clusters. Then it classifies the data set by dividing it into a given number of clusters of similar characteristics/preferences. The similarity is calculated using the distance between the two items. In this method, the distance is measured using a square Euclidean, Manhattan, Euclidean, or Cosine distance measure. This method is evaluated using the elbow method or silhouette analysis [22, 23, 24].

(2)(3)(4)(5)

where x1, y1, x2, y2 are the coordinates of the data points and () and () are the polar coordinates of x and y.

Naive Bayes

Naive Bayes [3] is an ML probabilistic algorithm that is based on the Bayes theorem. Such algorithms result in each pair of items or features being independent of each other. In Naive Bayes, the assumptions are that each feature provides an independent and equal part in the outcome. To start, the Bayes theorem is discussed below [26].

(6)

where P(X/Y) is the probability of X given that Y event has occurred, P(Y/X) is the probability of Y given that X event has occurred,

P(X) is the probability of event X, and

P(Y) is the probability of event Y.

The types of naive Bayes are: Bernoulli, Multinomial, and Gaussian naive Bayes.

Bernoulli naive Bayes: This is a binary algorithm that interprets whether a feature is present or not. It is used when there are binary function vectors (i.e., ones and zeroes). One of its applications is the bag of words model for text classification [27].

It follows the following rule:

(7)

where x and y are two events and i is a subevent of x.

Multinomial naive Bayes: Feature vector refers to the frequencies that are made using the multinomial distribution. It is used efficiently for working with texts in natural language processing.

Gaussian naive Bayes: Values associated with each feature vector are generated by Gaussian distribution or Normal distribution. If this is shown graphically, it results in a bell-shaped curve. The equation for this is as follows:

(8)

The steps involved in naive Bayes are written below:

1. The dataset is first preprocessed.

2. The fitting of Naive Bayes in the training data.

3. Predict the features of the test data.

4. Create the confusion matrix and get the accuracy of the model.

5. Try to visualize the result of the testing set.

The advantage of naive Bayes is that it is quick and precise for predictions. Such an approach also reduces the complexity of the computations. It can be used not only for one but also for problems with multiple feature classes. This algorithm works best if the variables are discrete and not continuous. The main disadvantage of naive Bayes is the assumption that features are independent of each other, which is not possible in real life. Moreover, if there is no training set for a particular class feature, this may result in a posterior probability of zero. This is known as the zero-frequency problem.

There are a variety of applications of naive Bayes. A major application of Naive Bayes lies in the recommendation system. If collaborative filtering and naive Bayes are both integrated into the recommendation system, it can predict through the unseen information regardless of preferences. As well, text classification is a popular application of naive Bayes. Applications of naive Bayes are real-time predictions and multiclass predictions for classification problems. It can also be used for facial recognition, medical testing, and weather forecasting.

Random Forest

The random forest algorithm [29] is a common supervised machine learning technique based on the ensemble learning concept. Ensemble learning is a method of combining various classifiers to improve model accuracy. In this algorithm, the dataset is split into several subsets and then contained in the same number of decision trees. Instead of depending on a decision tree, this algorithm takes an average of the predictions of all decision trees. This makes the outcome of the predictions more accurate.

The steps involved in implementing a random forest algorithm are given below:

1. The dataset is loaded and then preprocessed by splitting the data into a training and testing set.

2. The training and testing data are then feature scaled.

3. The training set is used to fit the random forest algorithm (defined as RandomForestClassifier). This is done by importing the sklearn library.

4. Prediction of the test result is made using a new prediction vector.

5. To conclude, a confusion matrix is created. This matrix gives the correct and incorrect predictions.

6. Visualization of the test result is done.

The main advantage of this algorithm is its versatility. It has increased predictability. So, this is a handy algorithm to use. It also overcomes the biggest problem of overfitting. It can handle a large dataset and also needs less time to train the dataset. The major drawback is that many decision trees can delay the algorithm and not function efficiently in the real world. It is used for both classification and regression, although it is not appropriate for regression.

There are various application domains of the random forest method. In banking, it is used for fraud detection, and loan risk identification, and various identifications and detections are performed based on banking services. In medicine, it is used to find the combination of medications and also to predict the risk and patterns of the disease. In commercialization, it can be used to predict stock prices and trends. It is also used in satellite imagery and object and multiclass detection.

Evaluation Methods

There are various methods used in the evaluation of machine learning methods. One of the commonly used methods is the absolute error and accuracy-based evaluation methods such as RMSE (Root mean squared Error), MSE (Mean square error), and MAE (Mean absolute error). There are decision support methods like precision, recall, F1-measure, and ROC (Receiver operating characteristic) curve. In addition, there are ranking-based evaluation methods, such as nDCG (Normalization of discounted cumulative gain), MRR (Mean reciprocal rank), mean precision, and Spearman rank correlation. Moreover, different metric evaluation methods assess performance based on prediction, decision, and ranking power. Examples of these metric-based approaches include coverage, popularity, novelty, diversity, and temporal evaluation. Finally, business sector metrics can be used to reach its objective. The above-mentioned algorithms will be evaluated using F1-measure, RMSE, and MAE.

F1. Measure

This accuracy measurement combines accuracy and recall and is also called the harmonic average of the model. This is used to measure the accuracy of the model.

The formula for the F1 measure is F1=2*P*R/(P+R), where P and R are the precision and recall of the model.

Precision: This measure, also known as the TP (True positives), is defined as the ration of TP to the sum of TP and FP (False positives).

Recall: This measure, also known as sensitivity, is defined as the ratio of the TP to the sum of TP and FN (False negatives).

(9)(10)

To avoid the least robustness of normal accuracy measurements, this measurement is preferred since it can take note of variations of different types of errors. The F1 measure is efficient whenever there is a presence of different costs of FP(False positives) and FN(False negatives). The F1 measurement can also be useful if there is an imbalance in the class feature numbers because, in such cases, the precision can be very misleading. The weakness of the F1 measurement is that the value calculated for one feature is independent of the other. In other words, it cannot compute the effectiveness of two features combined or based on each other's information. The applications for the F1 measurement include information retrieval in NLP (Natural Language Processing). This is most frequently used in search engine systems. In addition, it is most commonly used in binary classification systems.

RMSE (Root Mean Squared Error)

It is a performance measure of the ML models that are primarily calculated to see how well the model fits (i.e., less error, more accuracy). In other words, this is used to predict quantitative data. It is defined as:

(11)

In the above RMSE equation, is the original data and is the predicted data.

This measure is used because it is quite easy to distinguish. This makes it easier to work with methods such as gradient descent. This is also good for evaluating the standard deviation for distributing the errors generated.

RMSE has square errors, so even a small error can affect the value immensely, which allows us to ensure that the model yields as little error as possible. This means that an error of 10 will become 100 times worse than an error of 1. RMSE could become difficult to understand from an interpretation point of view as it contains square values, whereas MAE would be clear to understand due to absolute values.

MAE (Mean Absolute Error)

This measure is also used as an alternative to RMSE. MAE is the average of the absolute difference between the original data and the predicted data. If this absolute value is not taken, this will become the mean bias error (MBE).

To represent MAE mathematically:

(12)

In the above MAE equation, yj is the original data and yj^ is the predicted data. MAE is more stable than RMSE when the variation in frequency error distribution increases. This means that an error of 10 will be 10 times worse than an error of 1. MAE is generally preferable when scales of error are linear, whereas RMSE is preferable when scales of error are non-linear. MAE is not useful when no absolute value is required, in such cases, RMSE is preferable.

EXPERIMENTATION

Dataset

The BookCrossing dataset [34] is built by CAI-Nicolas Ziegler from Amazon Web Services. There are 270,000 books read by 90,000 users with 1.1 million reviews. The data consist of three tables which include information about ratings, books, and users. This data is downloaded from Kaggle. The rating dataset provides a list of book ratings given by the users. It includes 1,149,780 rating records containing 3 fields: userID, ISBN, and bookRating. The ratings are either explicitly expressed on a scale of 1 to 10 or implicitly expressed by zero. As shown in Fig. (7), the vast majority of ratings are 0 and these ratings are distributed very unevenly. The books dataset provides book information, which includes 271,360 book records containing 8 fields. First, 5 fields containing the content-based information: ISBN, Book-Title, Book-Author, Year-Of- Publication, Publisher, and the last 3 image-URL fields: Image-URL-S, Image-URL- M, Image-URL-L. These 3 different URL images are linked to the cover page of the books according to their size. The user dataset provides demographic information of users. It includes 278,858 user records and 3 fields: user id, Location, and Age. Fig. (8) shows that the majority of active users are youth between the ages of 20 and 30.

Fig. (7)) Rating Distribution of the books in dataset. Fig. (8)) Age Distribution of users in user-data.

Implementation

The book recommendation system has been done using item-based and user-based collaborative filtering experimented in python and compiled in Jupyter Notebook. After evaluating the RMSE scores of the user and the item, optimization of the book recommendation system is done by integrating various other algorithms, such as co- clustering, SVD, NMF, KNNbasic, KNNwithMeans, and KNNwithZScore models from the surprise library.

Result

In this article, the BookCrossing Dataset was implemented. It contains three tables. One table contains the user's information. The second table includes information on books. The final table includes the book routing information. Experimentation employed user-based and item-based collaborative filtering methods for the desired recommendation system. The RMSE score of these methods varied from 7 to 8. For improvement, the use of co-clustering, SVD, NMF, KNNbasic, KNNwithMeans, and KNNwithZScore models is done. The use of the above algorithms allowed a dramatic improvement of RMSE and MAE errors. The following table shows the RMSE value, the MAE value, and the F1 score for the implemented algorithms.

Discussion

By comparing the values obtained from the above analysis, the graphic display is shown below in Fig. (9) and Fig. (10). For a suitable algorithm, it is necessary to use a smaller RMSE and MAE measurement and a greater F1 measurement. Fig. (9). also represented that RMSE is higher than MAE. This is due to the differences mentioned earlier in Table 1. Thus, in the measure of errors, for comparison, the RMSE value is much better than the MAE value. As well, the F1 measurement is used for the confusion matrix Table 2 shows the comparison of different techniques in term of RMSE, MAE, and F1.

Fig. (9)) RMSE and MAE comparison of implemented algorithms. (Error(%) on y-axis and Algorithms on x-axis). Fig. (10)) F1 measure of implemented algorithms. (F1-measure(%) on y-axis and Algorithms on x-axis).
Table 1Difference between RMSE and MAE [33].MAERMSEIt doesn’t consider the sign of the input, if the input is negative it takes the positive value.It considers the sign of the input whether it is positive or negative.It is less biased towards large values. Thus, when it comes to a large error, it does not reflect the result of the algorithm.When it comes to large errors, it reflects in the result of the algorithm. Thus, it is much better than MAE.The MAE value is comparatively smaller as the sample size increases.RMSE is comparatively higher than MAE for increasing sample size.MAE restricts larger errors.RMSE does not restrict large errors.MAE is preferred where there is a proportion between overall performance and an increase in error.RMSE is preferred where the overall performance and the increase in error are disproportionate.
Table 2RMSE, MAE, and F1 measure of algorithms implemented on dataset.AlgorithmRMSEF1 measureMAECo-clustering1.83930.42891.4274SVD1.57260.44281.2046NMF2.47670.42022.0717KNNBasic1.94730.44341.5263KNNwithMeans1.79940.44041.3925KNNwithZScore1.79670.44021.3819

CONCLUSION

Hence, this work concludes that the SVD technique is the most preferred among the algorithms implemented. Fig. (9) shows that the NMF has a large RMSE and MAE and less F1 measurement compared to others. It further concludes that the NMF alone is not suitable for this dataset. Moreover, KNN (includes KNNBasic, KNNwithMeans, KNNwithZScore) is much better compared with NMF, primarily based on RMSE and MAE values. In addition, it concludes that the evaluation of the RMSE is much better than that of the MAE.

ACKNOWLEDGEMENT

We would like to express our gratitude to the Department of Computer Engineering at Dwarkadas J Sanghvi College of Engineering, who motivated us to dive into research and guided us when we faced any difficulty. Also, the assistance provided by our senior classmate Onkar Thorat is greatly appreciated.

References

[1]Silveira T., Zhang M., Lin X., Liu Y., Ma S.. How good your recommender system is? A survey on evaluations in recommendation., Int. J. Mach. Learn. Cybern..2019; 10[CrossRef][2]Shani G., Gunawardana A.. Tutorial on application-oriented evaluation of recommendation systems., AI Commun..2013; 26: 225-236. [CrossRef][3]Nguyen, Sang. , . “Model-Based Book Recommender Systems using Naïve Bayes enhanced with Optimal Feature Selection”, 217-222.[4]Isinkaye F.O., Folajimi Y.O., Ojokoh B.A.. , . 2015. “Recommendation systems: Principles, methods and evaluation”, Egyptian Informatics Journal, Volume 16, Issue 3, 2015, Pages 261-273,[5]Valdiviezo-Diaz P., Ortega F., Cobos E., Lara-Cabrera R.. A Collaborative Filtering Approach Based on Naïve Bayes Classifier., IEEE Access.2019; 7: 108581-108592. [CrossRef][6]Doshi S.. , . Brief on recommender systems. 2019.[7]Do, Minh-Phung Thi, D. V. Nguyen, and Loc Nguyen. "Model-based approach for collaborative filtering." In 6th International Conference on Information Technology for Education, pp. 217-228. 2010.[8]Laishram A.. , . Novelty in Recommender Systems. 2019.[9]Johnson W.. , . Recommender Systems with Apache Spark’s ALS function. 2016.[10]Ayse Yaman, CodeX, “Hybrid Recommender System-Netflix Prize Dataset”, Medium, Retrieved https://miro.medium.com/max/1370/0*PCZeW5TphSgtkIqm.png[11]Pontes B., Giráldez R., Aguilar-Ruiz J.S.. [CrossRef] “Biclustering on expression data: A review”, Journal of Biomedical Informatics, Vol. 57, pp. 163-180, ISSN 1532-0464[12]Gan X., Liew A.W-C., Yan H.. Discovering biclusters in gene expression data based on high-dimensional linear geometries., BMC Bioinformatics.2008; 9: 209. [CrossRef] [PubMed][13]Das, Joydeep & Mukherjee, Partha & Majumder, Subhashis & Gupta, Prosenjit, “Clustering-Based Recommender System Using Principles of Voting Theory”, Proceedings of 2014 International Conference on Contemporary Computing and Informatics, IC3I 2014. 10.1109/IC3I.2014.7019655, 2014.[14]Rege M., Dong M., Fotouhi F.. Co-clustering documents and words using bipartite isoperimetric graph partitioning, Proc. Int. Conf. Data Mining.2006: 532-541.[15]Chen Y., Dong M., Wan W.. Image co-clustering with multi-modality features and user feedbacks, Proc. Int. Conf. Multimedia.2009: 689-692.[16]Felzenszwalb P.F., Huttenlocher D.P.. Efficient graph-based image segmentation., Int. J. Comput. Vis..2004; 59(2): 167-181.[17]Luo J., Liu B., Cao B., Wang S.. Identifying miRNA-mRNA regulatory modules based on overlapping neighborhood expansion from multiple types of genomic data, Proc. Int. Conf. Intell. Comput.2016: 234-246.[18]Pio G., Ceci M., Loglisci C., D’Elia D., Malerba D.. Hierarchical and overlapping co-clustering of mRNA: miRNA interactions, Proc. Eur. Conf. Artif. Intell.2012: 654-659.[19]Hadrienj.github.io, . Deep Learning Book Series · 2.8 Singular Value Decomposition. 2019.[20]Andre L.V.P., Hruschka E.R.. [CrossRef] “Simultaneous co-clustering and learning to address the cold start problem in recommender systems”, Knowledge-Based Systems, Vol. 82, pp. 11-19, ISSN 0950-7051, 2015.[21]Aghdam, Hosseinzadeh. [CrossRef] Mehdi & Analoui, Morteza & Kabiri, Peyman. (2012). “Application of nonnegative matrix factorization in recommender systems”, 873-876, 2012.[22]Gupta S.. , . Top 5 Distance Similarity Measures implementation in Machine Learning. 2019.[23]2021.[24]Gupta S.. , . Top 5 Distance Similarity Measures implementation in Machine Learning. 2019.[25]Karbhari V.. , . What is a cosine similarity matrix?. 2020.[26]Ahadli T.. , . Naive Bayes Classifier: Bayesian Inference, Central Limit Theorem, Python/C++ Implementation. 2020.[27]Mutha N.. , . Bernoulli Naive Bayes.[28]Gandhi R.. , . Naive Bayes Classifier. 2018.[29]Ajesh A. , . “A random forest approach for rating-based recommender system”. pp. 1293-1297. 2016.[30]Al-Molegi A., Alsmadi I., Hassan N., Al-bashiri H.. Automatic Learning of Arabic Text Categorization., International Journal of Digital Contents and Applications..2015; 2: 1-16. [CrossRef][31]Dr Stylianos (Stelios) Kampakis, . Performance measures: RMSE and MAE.[32]JJ, . MAE and RMSE-Which Metric is Better?. 2016.[33]2021.[34]Cai-Nicolas Ziegler, . 2005.[35]Gipp B., Beel J., Hentschel C.. , . “Scienstein: A Research Paper Recommender System”, 2009.[36]Li T., Wang J., Chen H., Feng X., Ye F.. A NMF-based Collaborative Filtering Recommendation Algorithm, 6th World Congress on Intelligent Control and Automation.2006: 6082-6086.[37]Sahu S., Nautiyal A., Prasad M.. Machine Learning Algorithms for Recommender System - a comparative analysis., International Journal of Computer Applications Technology and Research..2017; 6: 97-100. [CrossRef][38]Portugal I., Alencar P., Cowan D.. The Use of Machine Learning Algorithms in Recommender Systems: A Systematic Review., Expert Syst. Appl..2015; 97[CrossRef][39]Nguyen, Sang. , . “Model-Based Book Recommender Systems using Naïve Bayes enhanced with Optimal Feature Selection”. pp. 217-222. 2019.[40]Gaudani H.. A Review Paper on Machine Learning Based Recommendation System., Development.2014; 2: 3955-3961.[41]Lampropoulos A., Tsihrintzis G.. Review of Previous Work Related to Recommender Systems., Intelligent Systems Reference Library..2015; 92: 13-30. [CrossRef][42]Nawrocka A., Kot A., Nawrocki M.. , . Application of machine learning in recommendation systems; 19th International Carpathian Control Conference (ICCC); 2018. p. 328.-331.[43]Babaee M., Tsoukalas S., Babaee M., Rigoll G., Datcu M.. [CrossRef