You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

27 September 2022

Multinomial Naive Bayesian Classifier Framework for Systematic Analysis of Smart IoT Devices

,
,
,
,
,
and
1
School of Computer Science, University of Petroleum and Energy Studies, Dehradun 248007, India
2
Software Engineering Department, College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi Arabia
3
Department of Management Information System, College of Business Administration, King Saud University, P.O. Box 28095, Riyadh 11437, Saudi Arabia
4
Department of Computer Science, Community College, King Saud University, P.O. Box 28095, Riyadh 11437, Saudi Arabia
This article belongs to the Section Internet of Things

Abstract

Businesses need to use sentiment analysis, powered by artificial intelligence and machine learning to forecast accurately whether or not consumers are satisfied with their offerings. This paper uses a deep learning model to analyze thousands of reviews of Amazon Alexa to predict customer sentiment. The proposed model can be directly applied to any company with an online presence to detect customer sentiment from their reviews automatically. This research aims to present a suitable method for analyzing the users’ reviews of Amazon Echo and categorizing them into positive or negative thoughts. A dataset containing reviews of 3150 users has been used in this research work. Initially, a word cloud of positive and negative reviews was plotted, which gave a lot of insight from the text data. After that, a deep learning model using a multinomial naive Bayesian classifier was built and trained using 80% of the dataset. Then the remaining 20% of the dataset was used to test the model. The proposed model gives 93% accuracy. The proposed model has also been compared with four models used in the same domain, outperforming three.

1. Introductions

Intelligent voice assistants, such as Microsoft’s Cortana, Apple’s Siri, Amazon’s Alexa, and Google’s Assistant, have grown significantly in popularity and use over the past several years. These intelligent voice assistants have changed how users interact with smartphones or computers. Individuals are using these intelligent voice assistants to give voice commands and get the appropriate information like daily news, weather reports, or fulfilling orders like playing media. Along with these uses, voice assistants are also used to perform basic tasks like setting timers or alarms and making phone calls. Nowadays, these voice assistants, especially Alexa, are also used in intelligent IoT-enabled devices to support voice control. The voice assistants are connected to the internet. Whenever a user gives any voice command, that command is sent to a central computing system for analysis. In the central computing system, the voice assistants analyze and translate the knowledge using natural language processing (NLP), and the voice assistant provides a proper response for that command. Recent advances in NLP have allowed voice assistants to generate meaningful responses rapidly [1].
With the help of artificial intelligence (AI), intelligent voice assistants can also detect or understand the user’s emotions and perform sentiment analysis. Sentiment analysis plays an important role when users share their feedback or experience regarding some product through voice assistants. Using sentiment analysis of voice assistants, commercial business companies can use insights to improve their products or services. It is essential for the voice assistant to accurately detect the sentiment in the users’ feedback or product review and analyze it to detect the user’s correct tone and mood. With the help of AI and NLP, the magnitude of the mood and tone of a user can be calculated, and a numerical score can be assigned to them. Depending on the outcome of the sentiment analysis, proper assistance can be provided to the user. As the popularity of intelligent voice assistants is increasing, the number of supported services is also growing rapidly. After using a product, most users like to share their experience about the product by writing reviews. This review not only helps the potential buyers but also helps the business companies make sound, impactful decisions about the product. So, it is essential to perform sentiment analysis on users’ reviews about the product and the services provided by that product. The product reviews are primarily available in text format. AI-based sentiment analysis can classify the products’ reviews into positive or negative categories by looking at the words used in the study. Generally, a positive product review contains words like good, easy, love, happy, and great, and a negative review contains words like disappointing, challenging, frustrating, harmful, waste, not, and annoy.
The authors have predicted customer sentiment from genuine Amazon Echo customer reviews using NLP in this research. The main objective of this research work is to indicate whether the users are happy or not with the Amazon Echo. If the customers are not satisfied with the product, then amazon can figure out the reason and help the users with proper assistance and update the product based on the reviews. In this research, a machine learning model has been built and trained to analyze thousands of studies of Amazon Echo to predict customer sentiments.
The remainder of this work is organized as follows:
  • The Related Work section highlights the relevant research done in the same area.
  • The Research Methodology section describes the step-by-step implementation of the method used in this research.
  • The Results and Comparison section compares the results with similar models in the same domain.
  • Conclusion sections conclude the investigation.

3. Research Methodology

In this paper, the authors have used Natural Language Processing techniques (NLP) to predict customer sentiment from genuine amazon Echo customer reviews. The dataset used for the implementation is https://www.kaggle.com/sid321axn/amazon-alexa-reviews/kernels (accessed on 15 July 2022). Amazon Echo Alexa is a type of virtual assistant. It can be placed anywhere in the home, and you can ask Alexa anything you want if you want it, for example, to book an appointment if you’re going to make a phone call, if you want to play music, you can say hey, Alexa do something. The main objective of this research is to try to predict customer sentiment through amazon reviews for the echo product. This research helps know whether the customers are happy or not with the product. Suppose the customers are not satisfied with the product. In that case, the manufacturer should know the reason behind customer dissatisfaction so that the organization can update the product based on the reviews. To analyze thousands of customer evaluations to forecast whether or not consumers are satisfied. This procedure might be carried out automatically with no involvement from humans. For analyzing a Smart device like Amazon Alexa, artificial intelligence is helpful. The authors have built a machine learning model to explore the IoT-based smart device.
Figure 2 illustrates the research methodology that the authors have followed in the implementation. The entire process is divided into four significant steps. The authors have implemented it in the python programming language. Initially, all the necessary and required libraries related to the implementation are imported, and exploratory data analysis is performed. In the second step, the data visualization of amazon Alexa reviews is performed, and word cloud is implemented. After that, the data cleaning and tokenization are done to improve the dataset quality. In this step, the dataset is prepared and ready for model development. In the final step, a machine learning model to analyze an IoT-based Smart Device is built and trained.
Figure 2. Research Methodology.
  • Step 1: Import the required libraries and dataset to perform exploratory data analysis
In the first step, the required libraries like NumPy, Pandas, Matplotlib, and Seaborn are imported for implementation. Seaborn is used for effective data visualization by plotting heatmap. Pandas are used to manipulate the data and to perform exploratory data analysis. The dataset contains the amazon Alexa reviews given by various users. The information related to the dataset is shown in Table 2 given below.
Table 2. Dataset Information.
As you can see in the table above that there are five columns in the dataset, out of the two columns rating and feedback are of integer type. Whereas, data, variation, and verified_reviews are of the object type. The rating column shows how many stars the customer had given to the product like a 1–5 rating. The date denotes the date on which the review is given. The variation shows the type of variant of Amazon Echo as if we have a Black Dot model, Charcoal Fabric model, Walnut Finish model, etc. The verified reviews are the actual reviews that we care about for implementation. That is the actual text that the authors have analyzed. Then finally, the last column is the feedback which is the sentiment. This is simply the customer sentiment, which is either zero or one. If some users say that–“I love my echo!” then that is positive sentiment and is represented by one whereas Zero represents the negative sentiment. Therefore, the rating and feedback column is important for visualization and training the machine learning model. As you can see in Table 3, there are 3150 rows in the dataset that has a non-null count. The statistical summary of the dataset is given in Table 3.
Table 3. Statistical Summary of the dataset.
As you can see in Table 3, the average rating is around 4.46 approximately the standard deviation is 1.06, which is the dispersion away from the mean is one approximately. The minimum review of the course is one star. The maximum review of the course is five stars. The statistical summary has 25%, 50%, and 75% as well. On average, the average ratings of around 3000 reviews. It stands around 4.46 approx. The statistical summary gives the overall mathematical view of the dataset for analyzing it.
  • Step 2: Perform the data visualization and plot the word cloud for amazon Alexa reviews
In this step, the data visualization is performed and the word cloud of amazon Alexa reviews is plotted to check positive and negative comments. As the authors have highlighted in the previous step, for building the deep neural network, we have focussed on two columns namely–rating and feedback. Therefore, the data visualization for these two columns specifically is important. Figure 3a represents the rating counts, as you can see that more than 2500 users out of 3150 users have given a five-star rating to the amazon Alexa product. Around 500 customers have given four stars rating and very few customers who are not happy gave it one star or two stars.
Figure 3. (a): Plot of rating counts, (b): Plot of feedback counts.
Figure 3b shows the plot of feedback counts, around 2800 customers are happy with the product and they have given positive feedback i.e., represented by one, and approximately 250 customers out of 3150 feedbacks are not happy with the product and they have given negative feedback represented by zero. As per the feedback plot, there are more satisfied customers as compared to unsatisfied customers.
From the data visualization, the authors have observed that there are zero missing or novel elements in the dataset. Moreover, the customers are also happy with the product as many customers gave it five-star reviews and many customers give it the sentiment of one indicating that they are happy with the product. After that, the authors performed some basic data exploration as shown in Figure 4, and calculated the length of the verified_review column. After observing the review’s length plot, we can conclude that there are a lot of reviews that are short in length. Maybe some customers are highly satisfied with the product and they wrote 2–3 words reviews and a few customers who are unhappy with the product, they wrote lengthy reviews.
Figure 4. Histogram plot for the length of the review.
After plotting the histogram of the length of the review, the authors have plotted the word cloud of positive and negative reviews. Word cloud is a really powerful visualization that will enable us to gain a lot of insight from text data. Here you can observe in Figure 5a, the wordcloud of positive reviews that were plotted by joining all positive reviews into one large string. All the sentences are clubbed together to form a large string and then the wordcloud was plotted.
Figure 5. (a) Wordcloud of positive reviews. (b) Wordcloud of negative reviews.
Here you can observe in Figure 6, the wordcloud of negative reviews that were plotted by joining all the negative reviews into one large string. All the sentences are clubbed together to form a large string and then the wordcloud was plotted.
Figure 6. Research Implemented.
The Algorithm 1 used for the creation of wordcloud of positive and negative reviews is given below:
Algorithm 1 Wordcloud Generation
Step 1: Convert all the positive/negative verified reviews into separate sentences
Step 2: Calculate the length of all those sentences
Step 3: Join all the verified positive/negative into a single large string
Step 4: Print that generated single large string
Step 5: Import the WordCloud python package and plot the positive/negative wordcloud of that joined single string
  • Step 3: Perform data cleaning and tokenization
In step 3, the data cleaning and tokenization are performed. Before building up the model, the data is cleaned first. Cleaning the data means that several stopwords, punctuation marks, exclamation marks, and common words are removed. The idea of cleaning the data is to convert text data into just a bunch of numbers. Once we have converted the text data into numbers we can use that data to train a machine learning model using these numbers to perform sentiment analysis. The pipeline is created to clean all the messages. Message cleaning means removing punctuation, and stopwords. The Algorithm 2 used to perform pipelining is given below:
Algorithm 2 Algo for Cleaning Messages
Defining cleaning_message(msg):
  no_punc = [If the char isn’t in the string, use char for the char in the message. punctuation]
  join_no_punc = ‘‘.join(no_punc)
  clean_join_no_punc = [If word.lower() is not in stopwords, split Test punc removed
join.split() word for word. words(‘english’)]
  return clean_join_no_punc
using sklearn.feature_extraction.text import CountVectorizer
vect = CountVect(analyz = cleaning_message)
reviews_countvect = vectorizer.fit_transform(reviews_df[‘verified_reviews’])
The sklearn library is used to vectorize and pipeline the dataset before feeding it to train the model. Creating a pipeline implies a series of perhaps functions or lines of code that can consume any text data and basically perform these three processes, remove punctuations and then remove stop words and then perform count vectorization.
  • Step 4: Build and train a Machine Learning Model to analyze a Smart IoT Device
In this step, we trained and tested the Naive Bayesian classifier model. In the previous step, we created a pipeline that can remove punctuations from text, remove stop words, and also perform count vectorization. And now we’re pretty much ready to go ahead and train our model. So the first step is to divide our data sets into training and testing. So in general when we train any machine learning model we use most of the data set perhaps around let’s say 80% of the data to train our model and then the remaining 20% we use them. And this is very important to make sure that our model is not overfitting the training data meaning the model still performs quite well even if the data has never been seen by the model before. And that’s why the actual true assessment of the model is on the testing data set which is a new, unique dataset for testing. In this step, initially, the entire dataset is divided into test and train data. Thereafter, the multinomial Naive Bayesian Classifier is implemented on the dataset. The test results are predicted in the form of a confusion matrix. Figure 6 shows the steps involved in the research implemented.

4. Results and Comparative Analysis

The model has been successfully trained and the performance of the model is visualized by plotting the confusion matrix as shown in Figure 7. The confusion matrix is simply a visual representation of our classifier model performance basically what we have here on the rows are our predictions. There are true classes in the confusion matrix, the true class represents the ground truth. The model predictions match the true class this means that the model is accurate. The model predictions were positive meaning that the model said based on what I’m reading right now from that customer review, it looks like the customer is happy for example, or maybe positive. If the true class was positive meaning if the actual customer was happy and left a positive review. The model predictions match what’s happening in real life and these are what we call through negatives basically in the diagonal elements here. If the model predicted negative and the true class was positive, we call that false-negative and we call that an error.
Figure 7. Confusion Matrix.
Table 4 shows the classification report for the predicted model. The classification report says that for Class-0, the precision is 0.81 and for class-1, the precision is around 0.95. The accuracy of our multinomial naive Bayesian classifier model came out to be 94%. The overall F1 score which is the harmonic mean between the precision and recall came out to be 0.99 for class-1.
Table 4. Classification Report.
For the comparative analysis, our proposed multinomial model is compared with the logistic regression also, the results obtained from logistic regression is shown in Table 5, it shows an accuracy of 93%, whereas our multinomial model obtained an accuracy of 94%.
Table 5. Classification Report of Logistic Regression.
To analyze the performance of our multinomial model, a comparative study is done with similar models in the same domain. After a comparative study as shown in Table 6, it has been observed that our model is the second-best model in the same domain. The metric used for the comparative study is the accuracy of the model.
Table 6. Comparative Analysis with similar approaches.
Text data from a number of sources, such as Facebook, Twitter, and Amazon, may be analysed and information can be extracted using a process known as sentiment analysis, sometimes known as opinion mining. In order for firms to actively work on enhancing their company strategy and gaining a complete understanding of the buyer’s viewpoint on their products, it is crucial. It comprises a computer study of a person’s buying habits, followed by a search for his opinions on a company’s commercial body. An activity, a person, a blog piece, or a shopping experience may all be examples of this entity. For this study, Amazon provided a dataset that included reviews of cameras, computers, mobile phones, tablets, televisions, and video surveillance. We used machine learning techniques to categorize positive and negative evaluations after pre-treatment. Machine Learning Techniques provide the greatest results for classifying Product Reviews, according to this research [22].
There are three types of reviews: favorable, impartial, and negative. This is useful not just for customers who want to read product reviews before buying, but also for businesses who want to see how the public reacts to their items. Using the Amazon API, we [23,24,25,26,27,28] were able to retrieve Amazon reviews. Machine learning classifiers have also been trained using unigrams and weighted unigrams. The results demonstrate that machine learning methods perform well on weighted unigrams, with SVM achieving the highest level of reliability [29,30,31,32,33,34]. The formulae used in the calculation of the classification report are given below:
A c c u r a c y = T P + T N   /   T P + T N + F P + F N
P r e c i s i o n = T P   /   T P + F P
R e c a l l = T P   /   T P + F N
F 1 s c o r e = 2   ×   P r e c i s i o n   ×   R e c a l l   /   P r e c i s i o n + R e c a l l
where TP = True Positive, TN = True Negative, FP = False Positive, FN = False Negative.
Numerous opinion mining approaches are used to extract the feelings hidden in the comments and reviews for a specific unlocked mobile. Additionally, a thorough analysis of sentiment classification methods is performed using the data set from mobile phone assessments [35,36,37,38,39,40,41,42]. The study’s [24] findings provide a comparison of eight distinct classifiers based on the assessment parameters of accuracy, recall, precision, and F-measure. Random Forest Classifiers are more efficient than other classifiers, however, LSTM and CNN are also more accurate.
Convolutional and max-pooling layers let the CNN model successfully retrieve higher-level data. With the help of the LSTM model, relationships between word sequences across time may be recorded. In this paper [25], we present a Hybrid CNN-LSTM Model, which combines LSTM with an extremely deep CNN model to solve the sentiment analysis problem. The proposed model additionally incorporates a rectified linear unit, dropout technology, and normalising to enhance accuracy. The proposed Hybrid CNN-LSTM Model outperforms conventional deep learning and ML methods in terms of precision, recall, f-measure, and accuracy.

5. Conclusions

NLP is used in the method of sentiment analysis to ascertain the attitude or sentiment of a text. A sentiment analysis algorithm may determine if a given text data is positive, negative, or neutral by deriving information from natural language and assigning it to a numerical score. There are several methods for developing or training a sentiment analysis model, and sentiment analysis is used by a variety of businesses to better understand their customers’ sentiments through reviews and social media conversations and make more timely and accurate business decisions. This study aims to offer a technique for assessing Amazon Echo customer evaluations and classifying them as good or negative. A dataset including the reviews of 3150 people was employed in this study. Initially, a word cloud of excellent and negative evaluations was produced, which provided a great deal of information from text data. After that, 80 percent of the dataset was used to build and train a machine learning model using a multinomial naive Bayesian classifier. The remaining 20% of the dataset was utilized to evaluate the model. The proposed model’s accuracy rate is 94%. The proposed model outscored three of the four models in the same area.

Author Contributions

Conceptualization: A.B., K.K., S.B. and S.D.; Methodology: A.B., K.K., S.B., S.D. and M.S.M.; Validation: A.B., K.K., S.B., M.S.M. and M.A.; formal analysis: A.B., S.D., K.K., S.B., M.S.M., M.A. and M.A.M.; investigation: A.B., K.K., S.B., M.A.M., M.S.M. and S.D.; resources: S.B., M.S.M., M.A. and S.D.; data curation: S.B., M.S.M., M.A. and M.A.M.; writing original draft preparation: A.B., K.K., S.B. and S.D.; writing review and editing: A.B., K.K., S.B., M.S.M. and M.A.; visualization: A.B., K.K., S.B., M.S.M. and M.A.; supervision: S.B., M.S.M. and M.A.M.; project administration: S.B.; funding acquisition: M.S.M. and M.A. All authors have read and agreed to the published version of the manuscript.

Funding

Research Supporting Project number (RSP2022R446), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The study did not report any data.

Acknowledgments

We deeply acknowledge King Saud University for supporting this study Research Supporting Project number (RSP2022R446), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hoy, M.B. Alexa, Siri, Cortana, and More: An Introduction to Voice Assistants. Med. Ref. Serv. Q. 2018, 37, 81–88. [Google Scholar] [CrossRef] [PubMed]
  2. Zwakman, D.S.; Pal, D.; Arpnikanondt, C. Usability Evaluation of Artificial Intelligence-Based Voice Assistants: The Case of Amazon Alexa. SN Comput. Sci. 2021, 2, 28. [Google Scholar] [CrossRef] [PubMed]
  3. Kaushik, K.; Singh, K. Security and trust in iot communications: Role and impact. In Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2020; Volume 989, pp. 791–798. [Google Scholar] [CrossRef]
  4. Natale, S.; Cooke, H. Browsing with Alexa: Interrogating the impact of voice assistants as web interfaces. Sage J. 2020, 43, 1000–1016. [Google Scholar] [CrossRef]
  5. Le Nguyen, M. Invited Talks Invited Talk #1 Deep Learning for NLG and Its Application For Chatbot System. In Proceedings of the 2018 5th NAFOSTED Conference on Information and Computer Science (NICS), Ho Chi Minh City, Vietnam, 23–24 November 2018; p. xxvii. [Google Scholar] [CrossRef]
  6. Lee, J.; Kim, D.; Sarikaya, R.; Kim, Y.B. Coupled Representation Learning for Domains, Intents and Slots in Spoken Language Understanding. In Proceedings of the 2018 IEEE Spoken Language Technology Workshop (SLT), Athens, Greece, 18–21 December 2018; pp. 714–719. [Google Scholar] [CrossRef]
  7. Zimmermann, A.; Schmidt, R.; Sandkuhl, K. Strategic Challenges for Platform-based Intelligent Assistants. Procedia Comput. Sci. 2020, 176, 966–975. [Google Scholar] [CrossRef]
  8. Saitulasi, K.; Deepa, N. Deep belief network and sentimental analysis for extracting on multi-variable features to predict stock market performance and accuracy. In Proceedings of the 2021 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 27–29 January 2021. [Google Scholar] [CrossRef]
  9. Garg, V.; Agarwal, B. Survey of Literature on Machine Intelligence and Deep learning for Smart Grid Applications. In Proceedings of the 2019 8th International Conference on Power Systems (ICPS), Rajasthan, India, 20–22 December 2019. [Google Scholar] [CrossRef]
  10. Mukherjee, S.; Nediyanchath, A.; Singh, A.; Prasan, V.; Gogoi, D.V.; Parmar, S.P.S. Intent Classification from Code Mixed Input for Virtual Assistants. In Proceedings of the 2021 IEEE 15th International Conference on Semantic Computing (ICSC), Laguna Hills, CA, USA, 27–29 January 2021; pp. 108–111. [Google Scholar] [CrossRef]
  11. Karri, S.P.R.; Kumar, B.S. Deep learning techniques for implementation of chatbots. In Proceedings of the 2020 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 22–24 January 2020. [Google Scholar] [CrossRef]
  12. Shah, A.; Jain, B.; Agrawal, B.; Jain, S.; Shim, S. Problem solving chatbot for data structures. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–10 January 2018; pp. 184–189. [Google Scholar] [CrossRef]
  13. Furey, E.; Blue, J. The Emotographic Iceberg: Modelling Deep Emotional Affects Utilizing Intelligent Assistants and the IoT. In Proceedings of the 2019 19th International Conference on Computational Science and Its Applications (ICCSA), St. Petersburg, Russia, 1–4 July 2019; pp. 175–180. [Google Scholar] [CrossRef]
  14. Parthasarathy, S.; Sundaram, S. Detecting expressions with multimodal transformers. In Proceedings of the 2021 IEEE Spoken Language Technology Workshop (SLT), Shenzhen, China, 19–22 January 2021; pp. 636–643. [Google Scholar] [CrossRef]
  15. Saibaba, C.H.M.H.; Waris, S.F.; Raju, S.H.; Sarma, V.; Jadala, V.C.; Prasad, C. Intelligent Voice Assistant by Using OpenCV Approach. In Proceedings of the 2021 Second International Conference on Electronics and Sustainable Communication Systems (ICESC), Coimbatore, India, 4–6 August 2021; pp. 1586–1593. [Google Scholar] [CrossRef]
  16. Singh, K.; Kaushik, K.; Ahatsham; Shahare, V. Role and Impact of Wearables in IoT Healthcare. In Advances in Intelligent Systems and Computing; Springr: Berlin/Heidelberg, Germany, 2020; Volume 1090, pp. 735–742. [Google Scholar] [CrossRef]
  17. Pal, S.; Hitchens, M.; Rabehaja, T.; Mukhopadhyay, S. Security Requirements for the Internet of Things: A Systematic Approach. Sensors 2020, 20, 5897. [Google Scholar] [CrossRef]
  18. Mahdavinejad, M.S.; Rezvan, M.; Barekatain, M.; Adibi, P.; Barnaghi, P.; Sheth, A.P. Machine learning for internet of things data analysis: A survey. Digit. Commun. Netw. 2018, 4, 161–175. [Google Scholar] [CrossRef]
  19. Ghazal, T.M.; Hassan, M.Q.; al-Sharida, M.T.; AlZoubi, H.M.; Ahmed, M.; Akbar, S.S.; Al-Kurdi, P.; Accor, I.A. IoT for Smart Cities: Machine Learning Approaches in Smart Healthcare—A Review. Future Internet 2021, 13, 218. [Google Scholar] [CrossRef]
  20. Zantalis, F.; Koulouras, G.; Karabetsos, S.; Kandris, D. A Review of Machine Learning and IoT in Smart Transportation. Future Internet 2019, 11, 94. [Google Scholar] [CrossRef]
  21. Jindal, M.; Gupta, J.; Bhushan, B. Machine learning methods for IoT and their Future Applications. In Proceedings of the 2019 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS), Greater Noida, India, 18–19 October 2019; pp. 430–434. [Google Scholar] [CrossRef]
  22. Jagdale, R.S.; Shirsat, V.S.; Deshmukh, S.N. Sentiment Analysis on Product Reviews Using Machine Learning Techniques. In Cognitive Informatics and Soft Computing; Springer: Berlin/Heidelberg, Germany, 2019; Volume 768, pp. 639–647. [Google Scholar] [CrossRef]
  23. Rathor, A.S.; Agarwal, A.; Dimri, P. Comparative Study of Machine Learning Approaches for Amazon Reviews. Procedia Comput. Sci. 2018, 132, 1552–1561. [Google Scholar] [CrossRef]
  24. Shaheen, M.; Awan, S.M.; Hussain, N.; Gondal, Z.A. Sentiment Analysis on Mobile Phone Reviews Using Supervised Learning Techniques. Int. J. Mod. Educ. Comput. Sci. 2019, 7, 32–43. [Google Scholar] [CrossRef]
  25. Rehman, A.U.; Malik, A.K.; Raza, B.; Ali, W. A Hybrid CNN-LSTM Model for Improving Accuracy of Movie Reviews Sentiment Analysis. Multimed. Tools Appl. 2019, 78, 26597–26613. [Google Scholar] [CrossRef]
  26. Bharany, S.; Badotra, S.; Sharma, S.; Rani, S.; Alazab, M.; Jhaveri, R.H.; Reddy Gadekallu, T. Energy efficient fault tolerance techniques in green cloud computing: A systematic survey and taxonomy. Sustain. Energy Technol. Assess. 2022, 53, 102613. [Google Scholar] [CrossRef]
  27. Bharany, S.; Sharma, S.; Badotra, S.; Khalaf, O.I.; Alotaibi, Y.; Alghamdi, S.; Alassery, F. Energy-Efficient Clustering Scheme for Flying Ad-Hoc Networks Using an Optimized LEACH Protocol. Energies 2021, 14, 6016. [Google Scholar] [CrossRef]
  28. Amin, F.; Lee, W.-K.; Mateen, A.; Hwang, S.O. Integration of Network science approaches and Data Science tools in the Internet of Things based Technologies. In Proceedings of the 2021 IEEE Region 10 Symposium (TENSYMP), Jeju, Korea, 23–25 August 2021; pp. 1–6. [Google Scholar] [CrossRef]
  29. Atlam, H.F.; Walters, R.J.; Wills, G.B. Fog computing and the internet of things: A review. Big Data Cogn. Comput. 2018, 2, 10. [Google Scholar] [CrossRef]
  30. Bharany, S.; Sharma, S.; Bhatia, S.; Rahmani, M.K.I.; Shuaib, M.; Lashari, S.A. Energy Efficient Clustering Protocol for FANETS Using Moth Flame Optimization. Sustainability 2022, 14, 6159. [Google Scholar] [CrossRef]
  31. Hameed, K.; Khan, A.; Ahmed, M.; Reddy, A.G.; Rathore, M.M. Towards a formally verified zero watermarking scheme for data integrity in the Internet of Things based-wireless sensor networks. Future Gener. Comput. Syst. 2018, 82, 274–289. [Google Scholar] [CrossRef]
  32. Bin Qaim, W.; Ometov, A.; Molinaro, A.; Lener, I.; Campolo, C.; Lohan, E.S.; Nurmi, J. Towards Energy Efficiency in the Internet of Wearable Things: A Systematic Review. IEEE Access 2020, 8, 175412–175435. [Google Scholar] [CrossRef]
  33. Bharany, S.; Sharma, S.; Khalaf, O.I.; Abdulsahib, G.M.; Al Humaimeedy, A.S.; Aldhyani, T.H.H.; Maashi, M.; Alkahtani, H. A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing. Sustainability 2022, 14, 6256. [Google Scholar] [CrossRef]
  34. Navas, R.E.; Cuppens, F.; Cuppens, N.B.; Toutain, L.; Papadopoulos, G.Z. MTD, Where Art Thou? A Systematic Review of Moving Target Defense Techniques for IoT. IEEE Internet Things J. 2020, 8, 7818–7832. [Google Scholar] [CrossRef]
  35. Bharany, S.; Kaur, K.; Badotra, S.; Rani, S.; Kavita; Wozniak, M.; Shafi, J.; Ijaz, M.F. Efficient Middleware for the Portability of PaaS Services Consuming Applications among Heterogeneous Clouds. Sensors 2022, 22, 5013. [Google Scholar] [CrossRef]
  36. Amin, F.; Ahmad, A.; Sang Choi, G.S. Towards Trust and Friendliness Approaches in the Social Internet of Things. Appl. Sci. 2019, 9, 166. [Google Scholar] [CrossRef]
  37. Valadares, D.C.G.; Will, N.C.; Caminha, J.; Perkusich, M.B.; Perkusich, A.; Gorgonio, K.C. Systematic Literature Review on the Use of Trusted Execution Environments to Protect Cloud/Fog-Based Internet of Things Applications. IEEE Access 2021, 9, 80953–80969. [Google Scholar] [CrossRef]
  38. Bharany, S.; Sharma, S. Intelligent Green Internet of Things: An Investigation. In Machine Learning, Blockchain, and Cyber Security in Smart Environments, 1st ed.; Chapman and Hall/CRC: London, UK, 2022; pp. 1–15. [Google Scholar]
  39. Bharany, S.; Sharma, S.; Frnda, J.; Shuaib, M.; Khalid, M.I.; Hussain, S.; Iqbal, J.; Ullah, S.S. Wildfire Monitoring Based on Energy Efficient Clustering Approach for FANETS. Drones 2022, 6, 193. [Google Scholar] [CrossRef]
  40. Amjad, A.; Azam, F.; Anwar, M.W.; Butt, W.H. A Systematic Review on the Data Interoperability of Application Layer Protocols in Industrial IoT. IEEE Access 2021, 9, 96528–96545. [Google Scholar] [CrossRef]
  41. Reilly, E.; Maloney, M.; Siegel, M.; Falco, G. A smart city IoT integrity-first communication protocol via an ethereum bloc chain light client. In Proceedings of the SERP4IoT, Colocated with the 44th ACM/IEEE International Conference on Software Engineering ICSE 2022, Marrakech, Morocco, 19 May 2022; pp. 15–19. [Google Scholar]
  42. El-Hajj, M.; Chamoun, M.; Fadlallah, A.; Serhrouchni, A. Analysis of authentication techniques in Internet of Things (IoT). In Proceedings of the 2017 1st Cyber Security in Networking Conference (CSNet), Rio de Janeiro, Brazil, 18–20 October 2017; pp. 1–13. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.