Intelligent Recommender System for Big Data Applications Based on the Random Neural Network †

Online market places make their profit based on their advertisements or sales commission while businesses have the commercial interest to rank higher on recommendations to attract more customers. Web users cannot be guaranteed that the products provided by recommender systems within Big Data are either exhaustive or relevant to their needs. This article analyses the product rank relevance provided by different commercial Big Data recommender systems (Grouplens film, Trip Advisor and Amazon); it also proposes an Intelligent Recommender System (IRS) based on the Random Neural Network; IRS acts as an interface between the customer and the different Recommender Systems that iteratively adapts to the perceived user relevance. In addition, a relevance metric that combines both relevance and rank is presented; this metric is used to validate and compare the performance of the proposed algorithm. On average, IRS outperforms the Big Data recommender systems after learning iteratively from its customer.


Introduction
Recommender systems were developed to address the search needs and relevant product selection within the Big Data from the biggest market place: the Internet.Whereas their benefit is the provision of a direct connection between customers and the desired products while reducing users' browsing time, any recommendation outcome will be influenced by a commercial interest as well as by the customers' own ambiguity in formulating their requests, queries or lack of product information.Online market places enable the trade of products provided by third party sellers while transactions are processed by the online market place operator; customers are provided with a service that search for products by their description or different properties such as department, brand, reviews or price.Another relevant application of Recommender Systems is travel industry: real time travel industry's information and services have been enabled by The Internet; customers directly purchase flight tickets, hotels and holiday packages via Web Pages and mobile applications where additional distribution costs have been eliminated due a shorter value chain.Although this direct purchase has benefited both customers and travel service providers, it has also created a reliance on Recommender Systems as products not presented on higher ranks within the suggestions may lose potential interested customers and revenue.In both examples, online market place and travel industry, product relevance is decided by ranking algorithms to transform information and products as visible or hidden to customers.Under this business model, Recommender System ranking algorithms could be tuned for a fee in order to artificially increase the rank of specific products whereas also product providers can be tempted to optimize the description of their products to manipulate the ranking algorithms.The main consequence to the final customer is that irrelevant products may be shown with a higher rank whereas relevant ones hidden at the very bottom of the recommended list.
In order to solve the described recommendation economic bias; this article presents a Big Data Intelligent Recommender System (IRS) based on the Random Neural Network as an interface between a customer's request and the recommender systems.IRS gets a request from the customer and obtains products from the recommender system data set allocating one neuron to each product dimension.The product relevance is evaluated by calculating an innovative cost formula that divides a request into a multidimensional vector and calculates its dimension terms with several relevance functions.
IRS adjusts and acquires the perceived customer's relevance and rearranges the obtained products based on the related centre point of the product's dimensions.IRS acquires product relevance on an iterative process where the customer assesses directly the presented products.Its performance is evaluated and compared against the Big Data Recommender Systems (Grouplens film, Trip Advisor and Amazon) with an innovative presented quality metric that combines relevance and rank.This article considers the term iteration as customer recommendation iterations rather than the learning algorithm iterations of the machine learning methods.Two different and independent learning algorithms are included, either Gradient Descent learns the related dimension centre or Reinforcement Learning updates the network weights recompensing related attributes while penalizing irrelevant ones.The research has compared IRS against the Big Data Recommender Systems with customer queries provided directly by validators.In addition, Gradient Descent and Reinforcement Learning algorithms have been independently analysed based on learning rate and product relevance.
The use of neural networks in recommender systems is described in Section 2. The Intelligent Recommender System mathematical model is defined in Section 3, Implementation in Section 4 and its validation against Big Data Recommender Systems is shown in Section 5. Finally, conclusions are shared in Section 6.

Recommender Systems
Customers interact with Recommender Systems via a user interface based on a Web portal or mobile app where a profiler extracts the customer properties based on feedback obtained from explicit and implicit methods; customers' interest on different products is predicted by ranking algorithms which provide a list of proposed items based on its calculated personalized relevance.Recommender system data architecture is mostly based on a database that stores and continuously updates item description and customer ratings (Figure 1).Due to the clustering and filtering services, Recommender Systems are extensively used within e-commerce [1] as they guide customers to discover new and rare products not found by themselves otherwise.manipulate the ranking algorithms.The main consequence to the final customer is that irrelevant products may be shown with a higher rank whereas relevant ones hidden at the very bottom of the recommended list.
In order to solve the described recommendation economic bias; this article presents a Big Data Intelligent Recommender System (IRS) based on the Random Neural Network as an interface between a customer's request and the recommender systems.IRS gets a request from the customer and obtains products from the recommender system data set allocating one neuron to each product dimension.The product relevance is evaluated by calculating an innovative cost formula that divides a request into a multidimensional vector and calculates its dimension terms with several relevance functions.
IRS adjusts and acquires the perceived customer's relevance and rearranges the obtained products based on the related centre point of the product's dimensions.IRS acquires product relevance on an iterative process where the customer assesses directly the presented products.Its performance is evaluated and compared against the Big Data Recommender Systems (Grouplens film, Trip Advisor and Amazon) with an innovative presented quality metric that combines relevance and rank.This article considers the term iteration as customer recommendation iterations rather than the learning algorithm iterations of the machine learning methods.Two different and independent learning algorithms are included, either Gradient Descent learns the related dimension centre or Reinforcement Learning updates the network weights recompensing related attributes while penalizing irrelevant ones.The research has compared IRS against the Big Data Recommender Systems with customer queries provided directly by validators.In addition, Gradient Descent and Reinforcement Learning algorithms have been independently analysed based on learning rate and product relevance.
The use of neural networks in recommender systems is described in section 2. The Intelligent Recommender System mathematical model is defined in Section 3, Implementation in Section 4 and its validation against Big Data Recommender Systems is shown in Section 5. Finally, conclusions are shared in Section 6.

Recommender Systems
Customers interact with Recommender Systems via a user interface based on a Web portal or mobile app where a profiler extracts the customer properties based on feedback obtained from explicit and implicit methods; customers' interest on different products is predicted by ranking algorithms which provide a list of proposed items based on its calculated personalized relevance.Recommender system data architecture is mostly based on a database that stores and continuously updates item description and customer ratings (Figure 1).Due to the clustering and filtering services, Recommender Systems are extensively used within e-commerce [1] as they guide customers to discover new and rare products not found by themselves otherwise.

Items
Recommender Algorithm

Web Portal
Item -User Database The model of a Recommender system (Figure 2) consists on:

Recommender System
• a set Q of N users, Q = {u1, u2, ..., uN} • a set I of M items, I = { i1, i2, ..., iM} The model of a Recommender system (Figure 2) consists on: • a set Q of N users, Q = {u 1 , u 2 , . . ., u N } • a set I of M items, I = { i 1 , i 2 , . . ., i M } • a rating matrix R, R = [r ij ] where i ∈ Q and j ∈ I The set of user-items {uN, iM} has an associated set of feature vector {xN} that represents customers with the different products assigned to them in the Content model.The relevance judgement pred(u,i) is a binary score, order or decision, based on the similarity between customers and products in the Collaborative model and weights in the Content model.

Recommender System Classification
Recommender systems can be mainly classified into two groups [2].Content-based recommender systems are founded on a product representation and a customer's preference profile; relevant products with similar features are identified based on the item and user properties without considering other customer's evaluations.Content-based approach suffer from some drawbacks [3] such as its incapability to suggest very different products and the need for the customer to rank numerous products beforehand in order to obtain useful recommendations.Collaborative recommender systems are based on the product feedback made by similar users and the customer's preceding ratings to other products; product suggestions are ranked based on the correlation between products and customers.Even though collaborative recommendation decreases the disadvantages from the Content-based option; it has also another issues such as the requirement of a significant volume of customer evaluation information in order to compute precise correlations and predictions, it also ignores on its calculations newly additional products or customers.Finally, Hybrid Recommender Systems combines both approaches to optimize their features while reducing their drawbacks.
Customers provide ratings of the various properties of a product as a rating vector in multicriteria ranking recommender systems [4], whereas products from several sources with locality collaborative filtering algorithms are suggested by cross domain recommender systems [5]; they The set of user-items {u N , i M } has an associated set of feature vector {x N } that represents customers with the different products assigned to them in the Content model.The relevance judgement pred(u,i) is a binary score, order or decision, based on the similarity between customers and products in the Collaborative model and weights in the Content model.

Recommender System Classification
Recommender systems can be mainly classified into two groups [2].Content-based recommender systems are founded on a product representation and a customer's preference profile; relevant products with similar features are identified based on the item and user properties without considering other customer's evaluations.Content-based approach suffer from some drawbacks [3] such as its incapability to suggest very different products and the need for the customer to rank numerous products beforehand in order to obtain useful recommendations.Collaborative recommender systems are based on the product feedback made by similar users and the customer's preceding ratings to other products; product suggestions are ranked based on the correlation between products and customers.Even though collaborative recommendation decreases the disadvantages from the Content-based option; it has also another issues such as the requirement of a significant volume of customer evaluation information in order to compute precise correlations and predictions, it also ignores on its calculations newly additional products or customers.Finally, Hybrid Recommender Systems combines both approaches to optimize their features while reducing their drawbacks.
Customers provide ratings of the various properties of a product as a rating vector in multi-criteria ranking recommender systems [4], whereas products from several sources with locality collaborative filtering algorithms are suggested by cross domain recommender systems [5]; they function by first modelling the typical similarity of the customer-product relation as a direct interconnected path and then analysing all the possible paths that connect customer or products in order to find new cross domain relations.
The evaluation of Recommender Systems in different e-commerce solutions is performed by various relevance metrics [6]; accuracy assessment measurements are divided into three different groups: classification of the correct decision making, rank of the right order and prediction of the difference between real and projected customer rankings.
A supplementary contribution to increase accuracy on Recommender Systems is based on the inclusion of Social network customer data [7]; social recommender systems using Collaborative Filtering methods are divided into two types: neighbourhood social approach based on social network graphs and matrix factorization approach based on the integration of customer item feedback history with customer to customer social data.

Personalized Information and Recommender Systems
Personal Recommender Systems are one of the most efficient tools to solve the information overload problem such as shopping online in an Internet marketplace.The common issue faced by different recommender systems is that both the customer and product information are dynamically changing [8]; in order to overcome this problem, an improved dynamic method based on updating local information is proposed to avoid the computational expensive process to update static information when an additional information is generated by the customer or product.A personalized recommendation methodology to increase the efficiency and value of suggestions when used to e-commerce applications consists on several of data mining methods such as Web usage learned from a clickstream, a decision tree induction that minimizes recommendation mistakes by providing recommendations only to users who are expected to purchase the recommended products, product taxonomy and finally association rule that chooses high e-commerce valuable items between possible recommendable products [9].
Models relevant to the application of personalized content services and their result on an improved customer experience and satisfaction can be identified as [10]: information overload theory that implies that customer satisfaction improves when the recommended content adjusts to customer interests, uses and gratifications theory focuses on incentives for access to information and customer involvement theory covers the customer's preferred content provided by a method that enables the costumer explicit participation.The Internet of Things (IoT) and Big Data sharing enable Personal Recommender Systems to model their customer models from external sources [11]; data from a variety of social Web services such as Facebook, LinkedIn can improve Personal Recommender System quality as it contains large amounts of personal information about customers.
Personal Recommender Systems can also be applied in traditional transport information systems [12], currently, customers must explicitly provide information related to both their profiles and journeys in order to receive a personalized response at additional extra effort from the customer in terms of search time; the proposed model identifies implicit customers' information and predicts their needs even if some data is missing by the use of an ontology that models more accurately semantic data.The inconsistency between the growth of tourism data and the struggle of tourists to retrieve tourism relevant information creates a business model for Tourism Information Recommender System [13]; the Recommender System solution merges the existing tourism information recommendation Websites where key relevance algorithms enable to take tourism decisions more accurately and promptly while permitting customers to expand their user experience of the tourism information service.
A customer model Web recommender system that predicts and personalizes music playlists is based on a hybrid similarity matching function that merges ontology-based semantic error measurements with collaborative filtering [14]; a personalized music playlist is dynamically produced from a collection of recommended playlists comprising the user most relevant tracks.A hybrid travel recommender system that suggests what suits the customer is based on the individual demographic information to make more intelligent and user acceptable recommendations [15]; the model merges demographic and collaborative filtering methods with the characteristics of content-based travel.

Neural Networks in Recommender System
Recommender Systems have also used Neural Networks as a solution to cluster customers or products into different groups or to detect and forecast customer rankings to different products.
Adaptive Resonance Theory (ART) is formed of a recognition threshold, a reset module and two layers of neurons: comparative and identification; the ART neural network is trained with an unsupervised learning algorithm.The k-separability algorithm has been applied in a collaborative filtering recommender system [16], the method is developed for each customer on several phases: a group of customers is clustered based on their similarity using an ART where the Singular Value Decomposition (SVD) matrix is determined with the k-separability algorithm applied to its neural configuration.The feed forward structured network has an input layer that consists of n neurons assigned to the customer rankings' matrix and an output layer that is formed on m neurons that feed the customer model.Customers are clustered into diverse categories using an ART model [17] where the vector that models the customer's features corresponds the neural network input layer and the output layers represent the applicable category.
A Self Organizing Map (SOM) is a neural network trained with unsupervised learning that reduces the dimensions of a quantified model for an input space.A Recommender System that applies a SOM with collaborative classification [18] is based on segmenting users by demographic characteristics in which users that correspond to the segments are clustered based on their item selection; then, the collaborative classification process is applied on the cluster allocated to the customer to recommend items where the product selection in each cluster is learned by the SOM, the input layer is the user cluster and the output layer is the group classification.The customer rating in a Recommender System application is calculated by a SOM [19] that forecasts the relevance of the unassessed products, the SOM is applied to classify the ranking group in order to fulfil a sparse ranking matrix.
Several solutions merge neural networks with collaborative filtering.A neural network recognizes implicit associations between customer preferences and potential relevant products to overcome the recommender system sparsity problem [20]; the patterns are applied to enhance personalized recommendations with collaborative filtering where the multilayer feed forward neural network structure is trained on every customer ranking vector, the output layer provides a prediction of the customer ranking vector that completes the unassessed products.A study of the hypothetic relationship between the students' final results and previous grades is founded on an Intelligent Recommender model framework [21] with a feed forward structure in a multi layered neural network that finds structures and relationships within the data based on a supervised learning process.
Any Machine Learning algorithm based on a feed forward neural network structure that consists on n input neurons, two hidden neurons and a single output neuron [22] could model collaborative filtering processes; the proposed algorithm reduces dimensionality using the Singular Value Decomposition (SVD) of an initial customer rating matrix that eliminates the need for users to rate common products with the purpose of predicting other user's interest.The input layer of the neural solution corresponds to a n singular vector that represents the averaged customer ranking whereas the output neuron characterizes the forecasted customer ranking.
Film recommendation systems have also been deployed with Neural Networks.A feed forward neural structure with only a hidden layer models a recommendation service that forecasts customer relevance to films based on its description, contextual data and feedback provided [23]; a TV program is characterized by a 24 dimensional attribute vector whereas the neural structure is formed of an input layer of five neurons (one for the day, three for the genre and one for the hour), a hidden layer and an output layer of two neurons: one for relevance and the other for irrelevance.A household viewer is assigned to its film ranking at a precise time by a neural network [24]; the input layer is formed of 68 nodes which correspond to the customer's temporal information and the 3 output nodes represents the different classifiers.An "Interior Desire System" solution [25] is based on the concept that if customers have similar browsing patterns, then they might have similar preference for explicit items; the back propagation neural structure organizes customers with matching navigation behaviour into clusters with comparable purchasing purpose patterns based on supervised learning.

Deep Learning in Recommender Systems
Recommender Systems algorithms have also been based on Deep Learning.A deep feature characterization [26] learns content data capturing the probability and implicit connection between products and customers where a Bayesian probabilistic model for the ranking matrix enables Collaborative classification.Customers and products are allocated into a vector space representation that improves the association among users and their preferred items by a Deep Learning method [27]; the method is expanded to acquire features of products from diverse groups and customer properties based on their search query log and Web browsing history, two diverse high dimensional sparse features are mapped by the neural structure into low dimensional dense features within a combined semantic space.
Deep neural networks are able to better generalize hidden feature interactions with low-dimensional dense embedding learned from the sparse characteristics, however this approach can over-generalize and recommend less relevant products when the customer-item combinations are sparse or with high ranks [28]; the proposed solution joins the benefits of wide linear learning that memorizes the sparse feature combinations using cross-product features with Deep Learning that generalizes previously unrevealed feature combinations.An approach that merges a collaborative filtering recommendation model with Deep Learning [29] applies a feature representation calculation using a quadric polynomial regression algorithm, which calculates more precisely the latent characteristics by improving the standard matrix factorization algorithm and the deep neural network method to predict the ranking values.
Most music is currently distributed, sold and consumed digitally making automatic music recommendation an increasingly relevant business opportunity in the last recent years.Deep convolutional neural networks are applied as a latent factor method for recommendation and prediction of the latent factors from music audio when they cannot be obtained from usage data [30], although a large semantic gap is found among the corresponding audio signal and the characteristics of a song that affect customer preference.Current content-based music recommendation methods normally extract traditional audio content characteristics such as Mel-frequency cepstral parameters before predicting customer preferences [31]; however, these standard features were not initially generated for music recommendation, a structure based on a deep belief network with a probabilistic graphical model automates the method that simultaneously learns characteristics from audio content and generates personalized recommendations.

Recommender Systems for Big Data
Big Data shall be processed and filtered before traditional Recommender Systems techniques can be applied to it due its extensive size and analysis time.A Recommendation System for the Big Data available on the Web in the structure of ratings, reviews, opinions, complains remarks, feedback and comments about any product is developed using Hadoop Framework [32]; the method uses a hybrid filtering technique based on numerical data such as ratings or ranks.Collaboratively Filtering does not provide enough scalability and accuracy when applied into Big Data Recommender Systems [33]; in order to overcome this issue, Big Data techniques such as association rule mining are inserted in three steps: Feature Extraction to generate the customer's product rating matrix, Low Dimension Latent Factor Matrix that uses the alternating least square method and Association Rule Mining Algorithm for generating multistage rule recommendations.
A hybrid recommendation technique for Big Data systems combines collaborative and content based filtering methods [34]; in addition, it also applies a product ranking and classification algorithm and social media opinion mining to analyse user feelings and hidden preferences from its posts.A personalized information recommendation method applies the measurement for context contribution and item correlation [35], the model improves the association rule algorithm and applies the Frequent-Pattern (FP) tree to increase the efficacy of the user behaviour pattern by mining in the Big Data structure.
A Big Data model for Recommender Systems uses social network data [36], the model incorporates factors such as human behaviour in the relationship and structure of the data in order to make better recommendations.A group centre recommender solution in the Cyber Physical Social Systems domain covers the detection of group activity, the analysis of ranking data for enhanced accuracy and the modelling of a group preference to enable context mining from several data sources [37], the proposed architecture is based on three components: group discovery for behavioural similarity, rating revision and multidimensional group preference modelling.
News platforms also present the challenges and prospects for applying the services of Big Data structures in recommendation services [38]; a news recommendation method is proposed to generate recommendation instructions at article level following the value model, the integration of the value and likelihood theme representations where the generation of recommendation instructions at theme level reduces the cold start issue.A Recommender System is based on the technical features that define a Big Data platform with a JSON wrapper data source layer, Hadoop knowledge based data management layer and rest API application layer [39]; the application has the capability to obtain information from distributed and diverse data sources, advanced data techniques, analysis, technologies, ability to provide personalized data to customers based on their preferences and enhanced information retrieval services.Matrix Factorization is an algorithm for collaborative filtering recommendation.A fast Tensor Factorization (TF) algorithm is applied for context-aware recommendation from implicit feedback [40]; for better accuracy, it includes all unobserved data to train a TF based context-aware recommender method by scrutinizing the matrix calculation of the closed form solution and memorizing the repetitive calculation to accelerate the process.
Matrix Factorization methods have confirmed a higher performance in data and image processing, however they have associated a greater computational complexity [41], a hybrid fusion collaborative recommendation solution maximizes feature engineering with feed forward neural networks to associate customer and product relationship into a dimensional feature space where product-customer similarity and the preferred products is optimized; Tensor Factorization (TF) improves generalization representing a multi-view data from customers that follow their purchasing and rating history.An improved recommendation model based on item-diversity adds customer interests and implicit feedback [42], it is based on the variance to the matrix factorization algorithm.A collaborative filtering application based on a deep learning method with a deep matrix factorization integrates any type of data and information [43]; two functions that transform features are incorporated to produce latent features of customers and products from diverse input data, in addition, an implicit feedback embedding transforms the implicit sparse and high-dimensional feedback data into a low dimension and real value vector that retains its primary factors.

Intelligent Recommender System Model
The Intelligent Recommender System is based on the Random Neural Network (RNN) [44][45][46].The RNN is a spiking recurrent stochastic structure that models more accurately how signals are transmitted in many biological neural structures propagating as impulses or spikes instead of analogue levels (Figure 3).Its mayor mathematical characteristics are the single neural network resolution for steady state and its product form equation.The RNN has been applied in several solutions such as the cognitive packet networks that search and select paths that meet specific Quality of Service metrics [47][48][49][50][51].
The RNN is composed of M neurons or nodes with a firing rate r(i), each node accepts excitatory and inhibitory spikes from outside sources which may be other nodes or sensors.The impulses are fired according to independent Poisson methods of rates λ + (i) for the excitatory impulse and λ − (i) for the inhibitory impulse respectively to neuron or node i transmitted in many biological neural structures propagating as impulses or spikes instead of analogue levels (Figure 3).Its mayor mathematical characteristics are the single neural network resolution for steady state and its product form equation.The RNN has been applied in several solutions such as the cognitive packet networks that search and select paths that meet specific Quality of Service metrics [47][48][49][50][51].
The RNN is composed of M neurons or nodes with a firing rate r(i), each node accepts excitatory and inhibitory spikes from outside sources which may be other nodes or sensors.The impulses are fired according to independent Poisson methods of rates λ + (i) for the excitatory impulse and λ -(i) for the inhibitory impulse respectively to neuron or node i Є {1,...M}.{1, . . .M}.
λi : Arrival rate of external inhibitory signal qi : Probability neuron excited ri : Rate neuron fires spikes p + : Probability excitatory signal neuron i -> neuron j p -: Probability inhibitory signal neuron i -> neuron j di di: Probability signal leaves network w + j,i = rip + i,j w -j,i = rip -i,j Non negative rates spike emission Excitatory / Inhibitory di + [p + i,j + p -i,j ] = 1 The state qi is the likelihood that the neuron i is active or excited which satisfies the non-linear equation: The examples used by animals when they search in a large unknown space were investigated and the role of teamwork was discussed [52]; these ideas were then applied to the search for information when N concurrent "search agents" with similar characteristics are being used [53] establishing that the total average search time can be reduced by aborting and re-starting a search process after a pre-determined time-out, if it has not been successful.The use of more concurrent agents can actually reduce the total energy costs, despite the increase in the number of agents, as well as the search times [54]; these results were confirmed in terms of large data sets distributed over large networks [55].

Recommendation Model
Recommendation of products or information needs the specification of three key components:  a N-dimensional universe of Z products to be recommended;  a request that defines the M-concepts or properties desired by a customer;  a process that finds and chooses X products from the universe Z, presenting the top Y products to the customer following a rule or method.
Each product in the universe is different from each other in a concept or property; on the other hand, the recommender model considers products to be different if they possess any distinct concept or meaning, even though if the customer may find them identical with respect to its request.
The Recommendation model considers the universe searched by the customer as a relation V formed of a set of Z N-tuples, V = {u1 , u2 … uZ}, where uo = (co1 , co2 … coN) and co are the N attributes for o=1, 2 ... Z.The relation V consists on N >> M attributes.CRt(m(t)) = (CRt(1), CRt(2), … , CRt(m(t))) defines a customer request where:  m(t) characterizes a variable M-dimension attribute vector with 1<M<N;  t represents the recommendation iteration with t>0.m(t) is variable to enable the addition or removal of attributes according to their relevance during the recommendation process.CRt(m(t)) obtains its parameters from the attributes that form the domain D(m(t)), where D represents the associated domain from which the universe V is created, therefore, D(m(t)) is a set of concepts or properties.
The reply R associated to the request CRt(m(t)) is a set of X N-tuples R = {u1 , u2 … uX} where ui = (ci1 , ci2 … ciN) and ci are the N attributes for i=1, 2 … X.The Intelligent Recommender System only The state q i is the likelihood that the neuron i is active or excited which satisfies the non-linear equation: The examples used by animals when they search in a large unknown space were investigated and the role of teamwork was discussed [52]; these ideas were then applied to the search for information when N concurrent "search agents" with similar characteristics are being used [53] establishing that the total average search time can be reduced by aborting and re-starting a search process after a pre-determined time-out, if it has not been successful.The use of more concurrent agents can actually reduce the total energy costs, despite the increase in the number of agents, as well as the search times [54]; these results were confirmed in terms of large data sets distributed over large networks [55].

Recommendation Model
Recommendation of products or information needs the specification of three key components: • a N-dimensional universe of Z products to be recommended; • a request that defines the M-concepts or properties desired by a customer; • a process that finds and chooses X products from the universe Z, presenting the top Y products to the customer following a rule or method.
Each product in the universe is different from each other in a concept or property; on the other hand, the recommender model considers products to be different if they possess any distinct concept or meaning, even though if the customer may find them identical with respect to its request.
The Recommendation model considers the universe searched by the customer as a relation V formed of a set of Z N-tuples, V = {u 1 , u 2 . . .u Z }, where u o = (c o1 , c o2 . . .c oN ) and c o are the N attributes for o = 1, 2 . . .Z.The relation V consists on N >> M attributes.CR t (m(t)) = (CR t (1), CR t (2), . . ., CR t (m(t))) defines a customer request where: • m(t) characterizes a variable M-dimension attribute vector with 1 < M < N; • t represents the recommendation iteration with t > 0. m(t) is variable to enable the addition or removal of attributes according to their relevance during the recommendation process.CR t (m(t)) obtains its parameters from the attributes that form the domain D(m(t)), where D represents the associated domain from which the universe V is created, therefore, D(m(t)) is a set of concepts or properties.The reply R associated to the request CR t (m(t)) is a set of X N-tuples R = {u 1 , u 2 . . .u X } where u i = (c i1 , c i2 . . .c iN ) and c i are the N attributes for i = 1, 2 . . .X.The Intelligent Recommender System only displays to the customer the top set of Y tuples that have the largest neuron potential value between the set of X tuples (Figure 4); the neuron potential is computed at every t iteration and characterizes the relevance of the M-tuple u i .The customer or the request itself are limited mainly by two mayor reasons: • the customer's absence of information or visibility to the entire attributes that create the universe V of products; • the customer's deficiency of defined understanding about the product it is searching.displays to the customer the top set of Y tuples that have the largest neuron potential value between the set of X tuples (Figure 4); the neuron potential is computed at every t iteration and characterizes the relevance of the M-tuple ui.The customer or the request itself are limited mainly by two mayor reasons:  the customer's absence of information or visibility to the entire attributes that create the universe V of products;  the customer's deficiency of defined understanding about the product it is searching.

Product Cost Formula
The Intelligent Recommender System model considers the universe V is created from the entire products that can be recommended.Each presented product is assigned to an N-tuple ui of the reply set R. The product relevance is calculated according to a product cost formula defined in this section.
The attributes the customer considers related are specified by the customer request CRt(m(t)) as a variable M-dimension vector where the dimension number of the attribute vector m(t) changes as the recommendation iteration t advances.The Intelligent Recommender System chooses products from the universe V assigning a N-tuple ui to every presented product generating a replay set R of X N-tuples.The product cost formula is applied to every product or N-tuple ui from the replay set R of X N-tuples where every ui is considered as a N-dimensional vector.The product cost formula is initially calculated using the related M attributes introduced by the customer on the request, CR1(m(1)) based on the domain D'(m(t)).The final Product Score, PS, is calculated by quantifying the relationship among the figures of the several attributes that it is formed: where PV is the Product Value that quantifies the product relevance and SF defines the Similarity Factor.The Similarity Factor (SF) recompenses products that have relevance scattered among their attributes where the dimensions or attributes of the customer request CRt(m(t)) are more related in the first positions rather than final ones: where SW[m] is the Similarity Weight, a M-dimension vector connected to the product or N-tuple ui and m is the attribute index from the customer request CRt(m(t)):

Product Cost Formula
The Intelligent Recommender System model considers the universe V is created from the entire products that can be recommended.Each presented product is assigned to an N-tuple u i of the reply set R. The product relevance is calculated according to a product cost formula defined in this section.
The attributes the customer considers related are specified by the customer request CR t (m(t)) as a variable M-dimension vector where the dimension number of the attribute vector m(t) changes as the recommendation iteration t advances.The Intelligent Recommender System chooses products from the universe V assigning a N-tuple u i to every presented product generating a replay set R of X N-tuples.The product cost formula is applied to every product or N-tuple u i from the replay set R of X N-tuples where every u i is considered as a N-dimensional vector.The product cost formula is initially calculated using the related M attributes introduced by the customer on the request, CR 1 (m(1)) based on the domain D'(m(t)).The final Product Score, PS, is calculated by quantifying the relationship among the figures of the several attributes that it is formed: where PV is the Product Value that quantifies the product relevance and SF defines the Similarity Factor.The Similarity Factor (SF) recompenses products that have relevance scattered among their attributes where the dimensions or attributes of the customer request CR t (m(t)) are more related in the first positions rather than final ones: where SW[m] is the Similarity Weight, a M-dimension vector connected to the product or N-tuple u i and m is the attribute index from the customer request CR t (m(t)): Score Dimension SD[m] is defined as a M-dimension vector that symbolizes the attribute values of every product or N-tuple u i connected the customer request CR t (m(t)).The Product Value (PV) is the addition of every dimension separate figure: where m is the attribute index from the customer request CR t (m(t)).Each dimension of the Score Dimension vector SD[m] is computed autonomously for every m-attribute figure that generates the customer request CR t (m(t)): The Intelligent Recommender System only considers three diverse types of domains of customer interest for this article: words, numbers and prices.The score S is assessed dependent if the domain of the attribute is a word, number or price.If the domain D(m) is a word, IRS computes the value of Score according to the equation: where the figure of WCR is 1 if the word of the m-attribute of the customer request CR t (m(t)) is present in the recommended product or it is assigned the value of 0 otherwise.NWR represents the amount of words in the recommended product.If the domain D(m) is a number, IRS selects the best score S from the numbers they are present in the recommended product that optimizes the product cost formula: where DVR represents the number of the m-attribute of the customer request CR t (m(t)), PVR defines the number within the product and NNP is the overall amount of numbers in the product.If the domain D(m) is a price, IRS selects the greatest Score from the prices in the product that optimizes the product cost formula: where DVR represents the price of the m-attribute of the customer request CR t (m(t)), PVR is the price in the product and NP is the overall amount of prices in the product.If the recommended product presents irrelevant price information, this is penalized by dividing the Score by the overall amount of prices in the product.The Position Parameter Factor (PPF) is created on the concept that an attribute displayed in the initial locations of the product recommendation is more related than if it is displayed at the final places: where NCP defines the amount of characters in the product and DVP corresponds to the location in the product where the dimension is displayed.The Relevance Parameter Factor (RPF) includes the customer's observation of relevance by recompensing the first attributes of the customer request CR t (m(t)) as very sought and penalizing the last attributes: where PDR is the location of the m-attribute of the customer request CR t (m(t)) and M is the overall amount of dimensions of the request vector CR t (m(t)).The Dimension Parameter Factor (DPF) includes the quantification of customer interest within the type of domains D(m(t)) by scoring higher on the domains that the customer has included more on the request: where NDT corresponds to the sum of the dimensions within equal domain (word, number or price) on the customer request CR t (m(t)) and M is the overall amount of dimensions of the request vector CR t (m(t)).The final Product Score figure (PS) is assigned to every N-tuple u i of the replay set R, this figure is applied by IRS to rearrange the reply set R of X N-tuples, displaying to the customer the top set of Y products which have the largest potential.

Customer Iteration
The customer, based on the replay set R evaluates the products provided and selects a subset of Q related products, I Q , of R. I Q is a set formed of Q N-tuples I Q = {u 1 , u 2 . . .u q }, u q is considered as a vector of N dimensions; u q = (c q1 , c q2 . . .c qN ) where c q are the N distinct attributes for q = 1, 2 . . .Q. Equally, the customer also chooses a subset of P unrelated products, I P of R, I P = {u 1 , u 2 . . .u p }, u p is defined as a vector of N dimensions; u p = (c p1 , c p2 . . .c pN ) where c p are the N distinct attributes for p = 1, 2 . . .P. Following the customer iteration, the Intelligent Recommender System shows to the customer an updated reply set R of Y N-tuples reordered to the lowest distance to the Related Centre and Unrelated Centre for the products selected.
The Related Centre Point is calculated according to the equation: where Q is the amount of related products chosen, m is the attribute index from the customer request CR t (m(t)) and SD q [m] the related Score Dimension vector to the product or N-tuple u q formed of c qm attributes.An equivalent formula is applied to the evaluation of the Unrelated Centre Point.
where P is the amount of unrelated products chosen, m the attribute index from the customer request CR t (m(t)) and SD p [m] the related Score Dimension vector to the product or M-tuple u p formed of c pm attributes.The Intelligent Recommender System rearranges the retrieved X set of N-tuples displaying only to the customer the top Y set of N-tuples based on the minimum value among the difference of their distances to the Related Centre Point and the Unrelated Centre Point: where LD is the Lowest Distance, RD represents the Related Distance to the product or N-tuple u i and UD corresponds to the Unrelated Distance to the same product.The Related Distance to every product or N-tuple u i is defined: where SD[m] corresponds to the Score Dimension vector of the product or M-tuple u i and RCP[m] represents the Related Centre Point spatial coordinates.Corresponding formula applies to the calculation of the Unrelated Distance: where SD[m] is the Score Dimension vector of the product or M-tuple u i and UCP[m] is the coordinate of the Unrelated Centre Point (Figure 5).An iterative recommendation process is presented that acquires and adjusts to the perceived customer relevance according to the dimensions or attributes the customer has initially selected on its request.An iterative recommendation process is presented that acquires and adjusts to the perceived customer relevance according to the dimensions or attributes the customer has initially selected on its request.

Attribute Learning
The replay set A to the request CR1(m(1)) is based on the M dimension customer request however products are generated of N dimensions consequently the subset of products the customer has evaluated as related might contain additional related hidden properties the customer did not thought on the initial request.The universe V is created from the domain D(n) or the N attributes which are considered as diverse and independent concepts from which the obtained set of X products are formed.The product cost formula is extended from the M attributes defined in the request CR1(m(1)) to the N attributes that generate the recommended products.The Score Dimension vector, SD[n], is now based on N-dimensions instead of M; a similar attribute extension is adjusted to calculate the Related Centre Point, RCP [n].
The request CR1(m(1)) consists on the M-Dimension vector inserted by the customer however the replay set R is formed of X N-tuples.The customer, according to the displayed replay set R, selects a subset of Q related products, IQ and a subset of P unrelated products, IP.IQ is considered as a set formed of Q M-tuples IQ = {u1 , u2 … uq} where uq is a vector of N dimensions; uq = (cq1 , cq2 … cqN) and cq are the N distinct attributes for q=1,2 ... Q.The N-dimension vector Dimension Average, DA[n], is the average figure of the n-attributes for the chosen related Q products: where Q represents the amount of related products selected, n the attribute index of the relation V and SDq[n] the related Score Dimension vector to the product or N-tuple uq created of cqN attributes.
ADV is considered as the Average Dimension Value of the N-dimension vector DA[n]:

Attribute Learning
The replay set A to the request CR 1 (m(1)) is based on the M dimension customer request however products are generated of N dimensions consequently the subset of products the customer has evaluated as related might contain additional related hidden properties the customer did not thought on the initial request.The universe V is created from the domain D(n) or the N attributes which are considered as diverse and independent concepts from which the obtained set of X products are formed.The product cost formula is extended from the M attributes defined in the request CR 1 (m(1)) to the N attributes that generate the recommended products.The Score Dimension vector, SD[n], is now based on N-dimensions instead of M; a similar attribute extension is adjusted to calculate the Related Centre Point, RCP [n].
The request CR 1 (m(1)) consists on the M-Dimension vector inserted by the customer however the replay set R is formed of X N-tuples.The customer, according to the displayed replay set R, selects a subset of Q related products, I Q and a subset of P unrelated products, I P .I Q is considered as a set formed of Q M-tuples I Q = {u 1 , u 2 . . .u q } where u q is a vector of N dimensions; u q = (c q1 , c q2 . . .c qN ) and c q are the N distinct attributes for q = 1, 2 . . .Q.The N-dimension vector Dimension Average, DA[n], is the average figure of the n-attributes for the chosen related Q products: where Q represents the amount of related products selected, n the attribute index of the relation V and SD q [n] the related Score Dimension vector to the product or N-tuple u q created of c qN attributes.ADV is considered as the Average Dimension Value of the N-dimension vector DA[n]: where N represents the overall amount of attributes that generate the relation V.The correlation vector σ[n] is the error between the dimension figures of every product with the average vector: where q is the amount of related products chosen, n the attribute index of the relation V and SD q [n] the related Score Dimension vector to the product or M-tuple u q generated of c qN attributes.C is defined as the average correlation figure of the N-dimensions of the vector σ[n]: where N is the overall amount of attributes that generate the relation V.An n-attribute is considered related if: • its Dimension Average figure DA[n] is greater than the average dimension ADV • its Correlation figure σ[n] is lower than the average correlation C.
On the following customer iteration, the request CR 2 (m(2)) is generated by the attributes the IRS has evaluated as related based on the above conditions.The replay to the request CR 2 (m(2)) is an updated set R of X N-tuples; this method reiterates until no new related products can be displayed to the customer (Figure 6).
where q is the amount of related products chosen, n the attribute index of the relation V and SDq[n] the related Score Dimension vector to the product or M-tuple uq generated of cqN attributes.C is defined as the average correlation figure of the N-dimensions of the vector σ[n]: where N is the overall amount of attributes that generate the relation V. On the following customer iteration, the request CR2(m(2)) is generated by the attributes the IRS has evaluated as related based on the above conditions.The replay to the request CR2(m(2)) is an updated set R of X N-tuples; this method reiterates until no new related products can be displayed to the customer (Figure 6).

Gradient Descent Learning
Gradient Descent learning adjusts to the appreciated customer relevance or comprehension of concepts by associating the attribute figures of every product; it analyses comparable concepts and rejects redundant ones.The IRS Gradient Descent learning algorithm consists on a recurrent structure

Gradient Descent Learning
Gradient Descent learning adjusts to the appreciated customer relevance or comprehension of concepts by associating the attribute figures of every product; it analyses comparable concepts and rejects redundant ones.The IRS Gradient Descent learning algorithm consists on a recurrent structure where the input nodes i = {i 1 , . . .,i Q } are the N-tuples u Q that correspond to the chosen related product subset I Q and the output nodes y = {y 1 , . . .,y Q } are fixed to the same values as the input nodes.IRS retrieves the network weights, computes the related attributes and finally rearranges the products following the lowest distance to the updated Related Centre Point that focus on the related attributes (Figure 7).

Reinforcement Learning
The customer choses the related product set IQ to provide the Reinforcement Learning exterior contact with the environment.Reinforcement Learning adjusts to the appreciated customer relevance by increasing the figure of related attributes and decreasing it for the unrelated ones.Reinforcement Learning adjusts the figures of the n attributes of the products, highlighting unseen related properties and ignoring unrelated concepts.The network weights are connected to the replay set R; W = R. IRS adjusts the network weights W by recompensing the product related attributes following: w q,n =c qn s-1 + c qn s-1 × c qn s-1 ∑ c qn s-1 N n=1 (22) where q is the product or M-tuple uQ created of cqm attributes, n is the product attribute index, M is the overall amount of attributes and s is the iteration value.IRS similarly adjusts the network weights by penalizing the product unrelated attributes according to: w q,n =c qn s-1 -c qn s-1 × c qn s-1 ∑ c qn s-1 N n=1 (23) IRS then calculates the potential value of every product following the adjusted network weights and rearranges them displaying to the customer the products that have the greater potential value (Figure 8).

Reinforcement Learning
The customer choses the related product set I Q to provide the Reinforcement Learning exterior contact with the environment.Reinforcement Learning adjusts to the appreciated customer relevance by increasing the figure of related attributes and decreasing it for the unrelated ones.Reinforcement Learning adjusts the figures of the n attributes of the products, highlighting unseen related properties and ignoring unrelated concepts.The network weights are connected to the replay set R; W = R. IRS adjusts the network weights W by recompensing the product related attributes following: where q is the product or M-tuple u Q created of c qm attributes, n is the product attribute index, M is the overall amount of attributes and s is the iteration value.IRS similarly adjusts the network weights by penalizing the product unrelated attributes according to: IRS then calculates the potential value of every product following the adjusted network weights and rearranges them displaying to the customer the products that have the greater potential value (Figure 8).

Implementation
The Intelligent Recommender System has been implemented in a server to enable online access to customers and validators.The main index page is developed using Java Server Pages (JSP) and Hyper Text Markup Language (HTML).Customers and validators access to the Web server via a Web browser and interact iteratively with the IRS invoking Java Servlets (Figure 9).The main software platform developed is Java Netbeans and the Web Server retrieval of data from the Internet is through Selenium Web-driver.The chosen server for this application is Apache Tomcat due its designed functionality to support Java.

Implementation
The Intelligent Recommender System has been implemented in a server to enable online access to customers and validators.The main index page is developed using Java Server Pages (JSP) and Hyper Text Markup Language (HTML).Customers and validators access to the Web server via a Web browser and interact iteratively with the IRS invoking Java Servlets (Figure 9).The main software platform developed is Java Netbeans and the Web Server retrieval of data from the Internet is through Selenium Web-driver.The chosen server for this application is Apache Tomcat due its designed functionality to support Java.

Implementation
The Intelligent Recommender System has been implemented in a server to enable online access to customers and validators.The main index page is developed using Java Server Pages (JSP) and Hyper Text Markup Language (HTML).Customers and validators access to the Web server via a Web browser and interact iteratively with the IRS invoking Java Servlets (Figure 9).The main software platform developed is Java Netbeans and the Web Server retrieval of data from the Internet is through Selenium Web-driver.The chosen server for this application is Apache Tomcat due its designed functionality to support Java.The Intelligent Recommender System acquires different dimensions values (Figure 10) from the user and the Learning Algorithm to use.The IRS has been programmed to retrieve Big Data from Recommender Systems data sets: GroupLens film, Trip Advisor and Amazon snippets.The process is transparent; IRS gets the request from the user and retrieves the Big Data without altering it.Once results have been retrieved from the data sets, our IRS applies the cost function to calculate their relevance.IRS finally reorders the results according to their relevance value.

Server
The Intelligent Recommender System provides to the customer a reordered list of products reranked based on a predetermined cost function using the Random Neural Network; the customer then ticks the relevant results and then the IRS provides to the customer with a reordered list based on the perceived customer relevance; this process repeats iteratively until the customer is satisfied with the recommendation or no additional products are found.

Quality
The following formula is proposed to measure recommender system quality; it is founded on the idea that a better recommender system suggests a list of more related products on higher order; in a set of P products, P is scored to the top product and 1 to the last product, the proposed quality figure is the addition of the order value for every selected product by the customer.The definition of Quality, Q, is represented as the following equation: where RPi is the rank of the product i with a figure of P if the product is in the top rank and 1 if the product is in the final rank.Z is the entire figure of products chosen by the customer.The best recommender system would provide the greatest Quality figure.Normalized quality, Q , is defined as the division of the quality, Q, by the optimum figure which it is when the customer evaluates as relevant all the products presented by the recommender system.In this circumstance, Z and P have identical figure: It is defined as the quality improvement between different iterations: where I is the Improvement, QRS represents the quality of the Recommender System and QR defines the quality reference.

Experiments
The Intelligent Recommender System is implemented to rearrange the products from three different and independent Big Data Recommender Systems data sets: GroupLens film, Trip Advisor The IRS has been programmed to retrieve Big Data from Recommender Systems data sets: GroupLens film, Trip Advisor and Amazon snippets.The process is transparent; IRS gets the request from the user and retrieves the Big Data without altering it.Once results have been retrieved from the data sets, our IRS applies the cost function to calculate their relevance.IRS finally reorders the results according to their relevance value.
The Intelligent Recommender System provides to the customer a reordered list of products re-ranked based on a predetermined cost function using the Random Neural Network; the customer then ticks the relevant results and then the IRS provides to the customer with a reordered list based on the perceived customer relevance; this process repeats iteratively until the customer is satisfied with the recommendation or no additional products are found.

Quality
The following formula is proposed to measure recommender system quality; it is founded on the idea that a better recommender system suggests a list of more related products on higher order; in a set of P products, P is scored to the top product and 1 to the last product, the proposed quality figure is the addition of the order value for every selected product by the customer.The definition of Quality, Q, is represented as the following equation: where RP i is the rank of the product i with a figure of P if the product is in the top rank and 1 if the product is in the final rank.Z is the entire figure of products chosen by the customer.The best recommender system would provide the greatest Quality figure.Normalized quality, Q, is defined as the division of the quality, Q, by the optimum figure which it is when the customer evaluates as relevant all the products presented by the recommender system.In this circumstance, Z and P have identical figure: It is defined as the quality improvement between different iterations: where I is the Improvement, QRS represents the quality of the Recommender System and QR defines the quality reference.

Experiments
The Intelligent Recommender System is implemented to rearrange the products from three different and independent Big Data Recommender Systems data sets: GroupLens film, Trip Advisor and Amazon.The IRS rearranges the products or films based on the updated product relevance computed by combining only the number of the selected dimensions.The larger the figure, the more relevant the product or film would be.IRS presents to the customer the first 20 products; the customer then selects the more relevant films or products on the first iteration; this ranking has been previously calculated by combining customer reviews to the same products and computing the average figure.Once IRS obtains customer feedback, it reorders the products based on the calculated relevance and presents to the customer a reordered list for further customer evaluation on the second iteration; this process repeats iteratively until the customer is satisfied with the recommendation or no additional products are found (Figure 11).This article considers the term iteration as customer recommendation iterations rather than the algorithm iterations of the machine learning methods.and Amazon.The IRS rearranges the products or films based on the updated product relevance computed by combining only the number of the selected dimensions.The larger the figure, the more relevant the product or film would be.IRS presents to the customer the first 20 products; the customer then selects the more relevant films or products on the first iteration; this ranking has been previously calculated by combining customer reviews to the same products and computing the average figure.
Once IRS obtains customer feedback, it reorders the products based on the calculated relevance and presents to the customer a reordered list for further customer evaluation on the second iteration; this process repeats iteratively until the customer is satisfied with the recommendation or no additional products are found (Figure 11).This article considers the term iteration as customer recommendation iterations rather than the algorithm iterations of the machine learning methods.In the evaluation, validators are asked to select relevant products, not to rank them, as they normally do when using a Recommender System therefore a product is considered as either relevant or irrelevant.Validators are 15 personal friends from Imperial College students, researchers and London young professionals degree educated.

GroupLens Film Data set
GroupLens is a research group in the Department of Computer Science and Engineering at the University of Minnesota.Since its creation in 1992 GroupLens' research projects have consisted on Recommender Systems, Online Communities, Mobile Technologies, Digital Libraries and Local Geographic Information Systems.The data set is based on a 5-star rating and genre tagging from MovieLens.It contains 21063128 ratings and 470509 tags across 27303 films.The data set was created by 229060 users between January-1995 and March-2015.
The data sets files are written as comma-separated values, csv extension.The files are encoded as UTF-8.The ratings data file consists of userId, movieId, rating and timestamp, whereas the film data file consists on movieId, title, genres.Genres are: Action, Adventure, Animation, Children, Comedy, Crime, Documentary, Drama, Fantasy, Film-Noir, Horror, Musical, Mystery, Romance, Sci-Fi, Thriller, War, Western and IMAX.The data set is processed by extracting the relevant information; movieId, rating title and genres.All the ratings from individual customers are combined to the same different products where the average value is the final product rating.Each film is considered as a multidimensional vector consisting on the different genres.The film relevance is the averaged film rating which is equally divided between the different genres it is classified.
The IRS is programmed to retrieve the user's relevant film genres and the type of learning (Gradient Descent or Reinforcement Learning).The IRS then reorders the films based on their relevance which is the combined value of the user selected genres only.The Intelligent Recommender System provides a reordered list of film titles to the user based on the higher values of the selected genres; the customer then selects the most relevant products with a higher overall rating; the IRS then continues the recommendation iterative process or provides with the final film titles.In the evaluation, validators are asked to select relevant products, not to rank them, as they normally do when using a Recommender System therefore a product is considered as either relevant or irrelevant.Validators are 15 personal friends from Imperial College students, researchers and London young professionals degree educated.

GroupLens Film Data Set
GroupLens is a research group in the Department of Computer Science and Engineering at the University of Minnesota.Since its creation in 1992 GroupLens' research projects have consisted on Recommender Systems, Online Communities, Mobile Technologies, Digital Libraries and Local Geographic Information Systems.The data set is based on a 5-star rating and genre tagging from MovieLens.It contains 21063128 ratings and 470509 tags across 27303 films.The data set was created by 229060 users between January-1995 and March-2015.
The data sets files are written as comma-separated values, csv extension.The files are encoded as UTF-8.The ratings data file consists of userId, movieId, rating and timestamp, whereas the film data file consists on movieId, title, genres.Genres are: Action, Adventure, Animation, Children, Comedy, Crime, Documentary, Drama, Fantasy, Film-Noir, Horror, Musical, Mystery, Romance, Sci-Fi, Thriller, War, Western and IMAX.The data set is processed by extracting the relevant information; movieId, rating title and genres.All the ratings from individual customers are combined to the same different products where the average value is the final product rating.Each film is considered as a multidimensional vector consisting on the different genres.The film relevance is the averaged film rating which is equally divided between the different genres it is classified.
The IRS is programmed to retrieve the user's relevant film genres and the type of learning (Gradient Descent or Reinforcement Learning).The IRS then reorders the films based on their relevance which is the combined value of the user selected genres only.The Intelligent Recommender System provides a reordered list of film titles to the user based on the higher values of the selected genres; the customer then selects the most relevant products with a higher overall rating; the IRS then continues the recommendation iterative process or provides with the final film titles.Gradient Descent and Reinforcement Learning have been validated independently for five different queries with ten recommendations in total.Tables 1 and 2 show the Quality for the different iterations with its associated 95% Confidence Range.I represents the improvement from IRS against the Recommender System; the second I is between IRS iterations 2 and 1 and finally the third I is between IRS iterations 3 and 2. Figure 12 shows the Quality for across the three different iterations for Gradient Descent (GD) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Figure 13 shows the associated improvement.Gradient Descent learns mostly on its first customer iteration whereas Reinforcement Learning improvement is dispersed between the different user iterations (Figures 12 and 13).The 95% Confidence Interval product regions overlaps.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent due its continued learning rate.The Hotel data set is the full reviews of hotels in 10 different cities (Dubai, Beijing, London, New York City, New Delhi, San Francisco, Shanghai, Montreal, Las Vegas, Chicago).There are approximately from 80 to 700 hotels in each city.The total number of reviews is approximately 259,000.The data set format is hotel id, hotel name, hotel url, street, city, state, country, post code, number of reviews, Cleanliness, Room, Service, Location, Value, Comfort and overall rating.Figure 13 shows the associated improvement.Gradient Descent learns mostly on its first customer iteration whereas Reinforcement Learning improvement is dispersed between the different user iterations (Figures 12 and 13).The 95% Confidence Interval product regions overlaps.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent due its continued learning rate.The Hotel data set is the full reviews of hotels in 10 different cities (Dubai, Beijing, London, New York City, New Delhi, San Francisco, Shanghai, Montreal, Las Vegas, Chicago).There are approximately from 80 to 700 hotels in each city.The total number of reviews is approximately 259,000.The data set format is hotel id, hotel name, hotel url, street, city, state, country, post code, number of reviews, Cleanliness, Room, Service, Location, Value, Comfort and overall rating.Gradient Descent learns mostly on its first customer iteration whereas Reinforcement Learning improvement is dispersed between the different user iterations (Figures 12 and 13).The 95% Confidence Interval product regions overlaps.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent due its continued learning rate.The Hotel data set is the full reviews of hotels in 10 different cities (Dubai, Beijing, London, New York City, New Delhi, San Francisco, Shanghai, Montreal, Las Vegas, Chicago).There are approximately from 80 to 700 hotels in each city.The total number of reviews is approximately 259,000.The data set format is hotel id, hotel name, hotel url, street, city, state, country, post code, number of reviews, Cleanliness, Room, Service, Location, Value, Comfort and overall rating.

Trip Advisor
The data set is processed by extracting its relevant information; the ratings from different years are combined into the same car type.The hotel and car data sets are joined into one where the average rating value is the final product (hotel or car) rating.Each product is a multidimensional vector consisting on the different variables the user can select; the product relevance is the product rating which it is equally divided between the different variables that is classified.
The IRS is programmed to retrieve the recommendation product (hotel or car) related attributes and the type of learning (Gradient Descent or Reinforcement Learning).The IRS then reorders the products based on their relevance which is the combined value of the user selected variables only.The IRS provides to the customer a reordered list of either hotels or cars re-ranked based on the higher values for the related attributes using the Random Neural Network; the customer then selects the most relevant products and the IRS continues the user feedback iterative process or provides the final hotel or car products.
Gradient Descent and Reinforcement Learning are evaluated independently for five different queries with ten recommendations in total for cars.Tables 3 and 4 show the Quality for different iterations with the associated 95% Confidence Range.I represents the improvement from IRS against the Recommender Systems; the second I is between IRS iterations 2 and 1 and finally the third I is between IRS iterations 3 and 2.    Figure 14 shows the Quality for across the three different iterations for Gradient Descent (GL) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Both Gradient Descent and Reinforcement learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 14 and 15).This learning rate is mostly because the initial quality was very high therefore difficult to improve.The 95% Confidence Interval is reduced and negligible when Reinforcement Learning reaches its optimum value; Reinforcement Learning manages to obtain a higher value of quality than Gradient Descent.
Gradient Descent and Reinforcement Learning are independently evaluated for five different queries with ten recommendations in total for hotels.Tables 5 and 6 show the Quality for different iterations with the associated 95% Confidence Range.I represents the improvement from IRS against  Figure 14 shows the Quality for across the three different iterations for Gradient Descent (GL) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Both Gradient Descent and Reinforcement learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 14 and 15).This learning rate is mostly because the initial quality was very high therefore difficult to improve.The 95% Confidence Interval is reduced and negligible when Reinforcement Learning reaches its optimum value; Reinforcement Learning manages to obtain a higher value of quality than Gradient Descent.
Gradient Descent and Reinforcement Learning are independently evaluated for five different queries with ten recommendations in total for hotels.Tables 5 and 6 show the Quality for different iterations with the associated 95% Confidence Range.I represents the improvement from IRS against Both Gradient Descent and Reinforcement learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 14 and 15).This learning rate is mostly because the initial quality was very high therefore difficult to improve.The 95% Confidence Interval is reduced and negligible when Reinforcement Learning reaches its optimum value; Reinforcement Learning manages to obtain a higher value of quality than Gradient Descent.
Gradient Descent and Reinforcement Learning are independently evaluated for five different queries with ten recommendations in total for hotels.Tables 5 and 6 show the Quality for different iterations with the associated 95% Confidence Range.I represents the improvement from IRS against the Recommender Systems; the second I is between IRS iterations 2 and 1 and finally the third I is between IRS iterations 3 and 2.     The obtained results are consisting with the previous evaluation; Both Gradient Descent and Reinforcement algorithms learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 16 and 17).Reinforcement Learning has a learning peak on its second iteration, this is because it provides the best scoring hotels in all cities whereas Gradient Descent still provides the best scoring hotels in the selected city only.

Amazon Data set
The Amazon data set contains product reviews and metadata, including 143.7 million reviews spanning from May 1996 to July 2014; the review data is an 18 GByte file.The subsets are: Books; Electronics; Movies and TV; CDs and Vinyl; Clothing; Shoes and Jewellery; Home and Kitchen; Kindle Store; Sports and Outdoors; Cell Phones and Accessories; Health and Personal Care; Toys and Games; Video Games; Tools and Home; Improvement; Beauty; Apps for Android; Office Products; Pet Supplies; Automotive; Grocery and Gourmet Food; Patio, Lawn and Garden; Baby; Digital Music; Musical Instruments and Amazon Instant Video.
Due the large processing time to analyse the entire data set, only the Films & TV subset is selected.Each Amazon subclass data set is form of two different sub sets:


The review set contains the reviewerID, productID, reviewer name, rating of the review, review text, rating of the product, summary of the review, time of the review.Figure 17 shows the associated improvement.The obtained results are consisting with the previous evaluation; Both Gradient Descent and Reinforcement algorithms learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 16 and 17).Reinforcement Learning has a learning peak on its second iteration, this is because it provides the best scoring hotels in all cities whereas Gradient Descent still provides the best scoring hotels in the selected city only.

Amazon Data set
The Amazon data set contains product reviews and metadata, including 143.7 million reviews spanning from May 1996 to July 2014; the review data is an 18 GByte file.The subsets are: Books; Electronics; Movies and TV; CDs and Vinyl; Clothing; Shoes and Jewellery; Home and Kitchen; Kindle Store; Sports and Outdoors; Cell Phones and Accessories; Health and Personal Care; Toys and Games; Video Games; Tools and Home; Improvement; Beauty; Apps for Android; Office Products; Pet Supplies; Automotive; Grocery and Gourmet Food; Patio, Lawn and Garden; Baby; Digital Music; Musical Instruments and Amazon Instant Video.
Due the large processing time to analyse the entire data set, only the Films & TV subset is selected.Each Amazon subclass data set is form of two different sub sets:


The review set contains the reviewerID, productID, reviewer name, rating of the review, review text, rating of the product, summary of the review, time of the review.The obtained results are consisting with the previous evaluation; Both Gradient Descent and Reinforcement algorithms learn mostly on their customer first iteration with a residual learning on their successive iterations (Figures 16 and 17).Reinforcement Learning has a learning peak on its second iteration, this is because it provides the best scoring hotels in all cities whereas Gradient Descent still provides the best scoring hotels in the selected city only.

Amazon Data Set
The Amazon data set contains product reviews and metadata, including 143.7 million reviews spanning from May 1996 to July 2014; the review data is an 18 GByte file.The subsets are: Books; Electronics; Movies and TV; CDs and Vinyl; Clothing; Shoes and Jewellery; Home and Kitchen; Kindle Store; Sports and Outdoors; Cell Phones and Accessories; Health and Personal Care; Toys and Games; Video Games; Tools and Home; Improvement; Beauty; Apps for Android; Office Products; Pet Supplies; Automotive; Grocery and Gourmet Food; Patio, Lawn and Garden; Baby; Digital Music; Musical Instruments and Amazon Instant Video.
Due the large processing time to analyse the entire data set, only the Films & TV subset is selected.Each Amazon subclass data set is form of two different sub sets:

•
The review set contains the reviewerID, productID, reviewer name, rating of the review, review text, rating of the product, summary of the review, time of the review.

•
The metadata set contains productID, name of the product, price, url of the product image, related products (also bought, also viewed, bought together, buy after viewing), sales rank information, brand name and the list of categories the product belongs to.
The data set is processed by extracting its relevant information and combining all the ratings from individual customers to the same products; the average value is the final product rating.Each product is a multidimensional vector consisting of the different properties or dimensions defined by the customer on its request.The Intelligent Recommender System provides to the customer a reordered list of products re-ranked; the customer then selects the most relevant products and the IRS continues the customer feedback iterative process or provides the final products.
Gradient Descent and Reinforcement Learning are independently validated for five different queries with ten recommendations in total.Tables 7 and 8 show the Quality for different iterations with its associated 95% Confidence Range.I represents the improvement from IRS against the Recommender System; the second I is between IRS iterations 2 and 1 and finally the third I is between IRS iterations 3 and 2.     Figure 18 shows the Quality for across the three different iterations for Gradient Descent (GD) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Gradient Descent learns predominantly during its first customer iteration whereas Reinforcement Learning improvement is more dispersed among the different user iterations (Figures 18 and 19).The 95% Confidence Interval is greater in the amazon data set with a significant product overlap.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent because of its continued learning speed.The Related Centre learned by Gradient Descent has defocused on its third iteration producing a decrement on Quality.
Overall, Gradient Descent predominantly acquires relevance on its first customer iteration with a decreasing learning speed in further stages.Reinforcement algorithm learns gradually with a learning speed dispersed within the customer iteration stages until it reaches its largest quality.IRS Figure 19 shows the associated improvement.Figure 18 shows the Quality for across the three different iterations for Gradient Descent (GD) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Gradient Descent learns predominantly during its first customer iteration whereas Reinforcement Learning improvement is more dispersed among the different user iterations (Figures 18 and 19).The 95% Confidence Interval is greater in the amazon data set with a significant product overlap.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent because of its continued learning speed.The Related Centre learned by Gradient Descent has defocused on its third iteration producing a decrement on Quality.
Overall, Gradient Descent predominantly acquires relevance on its first customer iteration with a decreasing learning speed in further stages.Reinforcement algorithm learns gradually with a learning speed dispersed within the customer iteration stages until it reaches its largest quality.IRS Gradient Descent learns predominantly during its first customer iteration whereas Reinforcement Learning improvement is more dispersed among the different user iterations (Figures 18 and 19).The 95% Confidence Interval is greater in the amazon data set with a significant product overlap.Gradient Descent outperforms Reinforcement Learning in the first iteration however Reinforcement Learning overtakes Gradient Descent because of its continued learning speed.The Related Centre learned by Gradient Descent has defocused on its third iteration producing a decrement on Quality.
Overall, Gradient Descent predominantly acquires relevance on its first customer iteration with a decreasing learning speed in further stages.Reinforcement algorithm learns gradually with a learning speed dispersed within the customer iteration stages until it reaches its largest quality.IRS algorithm provides very high quality with the Trip Advisor data set at its first customer iteration stage therefore the learning speed decreased significantly in further iterations, however it maintained equal learning pattern: Gradient Descent learns better on its first customer iteration while Reinforcement Learning progressively maintains its learning speed.

Discussion
A novel approach to recommendation systems in the Big Data is proposed where the customer recursively trains the neural network while searching for relevant products.An innovative method is defined: the use of the Random Neural Network as a biological inspired algorithm to quantify customer relevance and product ranking based on a cost formula.
The Intelligent Recommender System adjusts iteratively to the customer and acquires relevance based from customer feedback improving its quality in the initial iteration.Reinforcement Learning algorithm learns more than Gradient Descent.Although Gradient Descent presents larger quality on the initial iteration; Reinforcement Learning outpaces on the next iteration due its greater learning speed where both algorithms learn negligible in their final iteration.If only one customer iteration is possible, Gradient Descent shall be chosen learning algorithm; however, if two customer interactions are enabled, Reinforcement Learning is the best choice.More than three customer iterations are not recommended due to their learning is minimal.
The main challenge presented during this research has been the retrieval of evaluation data; the difficulty to obtain numerous user recommendations and experiments (in the order of 10 s rather than 1000 s) has increased our 95% Confidence Interval with regions of overlap products.The results are shown as "proof of concept" rather than statistically significant or small confidence intervals however statistical significance to demonstrate estimation and hypothesis of the research data has been also reported within the evaluation sections.
An n-attribute is considered related if:  its Dimension Average figure DA[n] is greater than the average dimension ADV  its Correlation figure σ[n] is lower than the average correlation C.
Data set Trip Advisor data set has been obtained from University of California-Irvine (UCI), Machine Learning repository, Centre for Machine Learning and Intelligent Systems.It is formed on two data sets; car and hotels reviews.The car data set is the full review of cars for model years 2007, 2008 and 2009.There are approximately from 140 to 250 different cars for each model year.The total number of reviews is approximately 42,230, (Year 2007: 18,903 reviews, Year 2008: 15,438 reviews and Year 2009: 7,947 reviews).The data set format is car model, year, number of reviews, Fuel, Interior, Exterior, Build, Performance, Comfort, Reliability, Fun and overall rating.

Figure 14
Figure14shows the Quality for across the three different iterations for Gradient Descent (GL) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.

Figure 14 .
Figure 14.Recommender Evaluation-Trip Advisor Car Database Figure 15 shows the associated improvement.

Figure 14 .
Figure 14.Recommender Evaluation-Trip Advisor Car Database Figure 15 shows the associated improvement.

Figure 14 .
Figure 14.Recommender Evaluation-Trip Advisor Car Database Figure 15 shows the associated improvement.

Figure 16 29 Figure 16 .
Figure 16 shows the Quality for across the three different iterations for Gradient Descent (GL) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.Big Data Cogn.Comput. 2019, 3, 1523 of 29

Figure 17
Figure17shows the associated improvement.

Figure 18
Figure18shows the Quality for across the three different iterations for Gradient Descent (GD) and Reinforcement Learning (RL) Algorithms with the 95% Confidence Interval that corresponds to Q ± 95%CR.

Figure 19
Figure19shows the associated improvement.

Figure 19
Figure19shows the associated improvement.
5.2.2.Trip Advisor Data set Trip Advisor data set has been obtained from University of California-Irvine (UCI), Machine Learning repository, Centre for Machine Learning and Intelligent Systems.It is formed on two data sets; car and hotels reviews.The car data set is the full review of cars for model years 2007, 2008 and 2009.There are approximately from 140 to 250 different cars for each model year.The total number of reviews is approximately 42,230, (Year 2007: 18,903 reviews, Year 2008: 15,438 reviews and Year 2009: 7,947 reviews).The data set format is car model, year, number of reviews, Fuel, Interior, Exterior, Build, Performance, Comfort, Reliability, Fun and overall rating.
5.2.2.Trip Advisor Data Set Trip Advisor data set has been obtained from University of California-Irvine (UCI), Machine Learning repository, Centre for Machine Learning and Intelligent Systems.It is formed on two data sets; car and hotels reviews.The car data set is the full review of cars for model years 2007, 2008 and 2009.There are approximately from 140 to 250 different cars for each model year.The total number of reviews is approximately 42,230, (Year 2007: 18,903 reviews, Year 2008: 15,438 reviews and Year 2009: 7947 reviews).The data set format is car model, year, number of reviews, Fuel, Interior, Exterior, Build, Performance, Comfort, Reliability, Fun and overall rating.