You are currently on the new version of our website. Access the old version .
Future InternetFuture Internet
  • Article
  • Open Access

6 March 2025

A Neural-Symbolic Approach to Extract Trust Patterns in IoT Scenarios

,
and
1
Department of Mathematics and Informatics, University of Catania, 95125 Catania, Italy
2
Department DIIES, University “Mediterranea” of Reggio Calabria, 89122 Reggio Calabria, Italy
3
Department of Psychology, University of Milan-Bicocca, 20126 Milan, Italy
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Joint Design and Integration in Smart IoT Systems

Abstract

Trust and reputation relationships among objects represent key aspects of smart IoT object communities with social characteristics. In this context, several trustworthiness models have been presented in the literature that could be applied to IoT scenarios; however, most of these approaches use scalar measures to represent different dimensions of trust, which are then integrated into a single global trustworthiness value. Nevertheless, this scalar approach within the IoT context holds a few limitations that emphasize the need for models that can capture complex trust relationships beyond vector-based representations. To overcome these limitations, we already proposed a novel trust model where the trust perceived by one object with respect to another is represented by a directed, weighted graph. In this model, called T-pattern, the vertices represent individual trust dimensions, and the arcs capture the relationships between these dimensions. This model allows the IoT community to represent scenarios where an object may lack direct knowledge of a particular trust dimension, such as reliability, but can infer it from another dimension, like honesty. The proposed model can represent trust structures of the type described, where multiple trust dimensions are interdependent. This work represents a further contribution by presenting the first real implementation of the T-pattern model, where a neural-symbolic approach has been adopted as inference engine. We performed experiments that demonstrate the capability in inferring trust of both the T-pattern and this specific implementation.

1. Introduction

Today, the paradigm of programming smart objects is evolving to incorporate social aspects, driven by the necessity for these objects to interact dynamically to perform complex tasks. These tasks may involve exchanging services, making requests, or negotiating contracts [1,2,3]. The representation of social interactions between software entities has naturally emerged within the field of cooperative smart objects [4,5], where leveraging social characteristics enables collaborative behaviors. Various approaches have been introduced to enhance collaboration among smart objects within their communities [6]. Additionally, in competitive environments such as e-commerce, several models have been developed [7,8] to equip smart objects with social capabilities that improve their ability to achieve specific goals. Across these models, the need to effectively represent trustworthiness among community members remains a consistent requirement. This necessity has led to the development of trust-based models that address different dimensions of trust relationships. Such models are typically designed to enhance smart object performance, particularly in IoT environments, where objects operate within vast networks and must constantly share information.
When two smart objects interact, the one requesting a service is termed the trustor, while the one providing it is the trustee. A trust relationship between two smart objects often spans multiple dimensions. The term dimension refers to the specific perspective used to evaluate the interaction. Common trust dimensions include attributes such as competence, honesty, security, reliability, and expertise, alongside others that depend on the specific context. These dimensions represent subjective measures that each smart object independently calculates based on its perception of its surroundings. However, in some cases, an additional dimension—reputation—is necessary. This is where the community collectively assigns a trust value to a trustee. Reputation becomes particularly relevant when a smart object o 1 lacks direct knowledge of another smart object o 2 and must instead rely on the community’s collective assessment of o 2 ’s trustworthiness.
Various trust and reputation models have been proposed for multi-smart object communities [9,10]. Most of these approaches quantify trust using scalar measures, integrating them into synthetic trustworthiness indicators. However, a key limitation of these models is their reliance on simple scalar values, often structured as trust vectors. For example, the trust of a smart object o 1 toward another object o 2 might be represented by the vector t r u s t o 1 , o 2 = [ h o n e s t y , r e l i a b i l i t y , e x p e r t i s e , ] . These trust vector elements are often interdependent rather than fully independent. For instance, a smart object might infer honesty from reliability, assuming that a reliable partner is also likely to be honest to some degree, factoring in both internal and external conditions. Our study does not introduce new quantitative trust measures but instead presents a model capable of representing scenarios where a smart object lacks direct knowledge of a trust dimension, such as reliability, and infers it from another, like honesty. The proposed model captures the interdependencies among multiple trust dimensions. To address this, we propose a trust and reputation model for social smart object communities, where the trust a smart object o 1 perceives toward another o 2 is not represented as a vector of independent values. Instead, it is modeled as a directed, weighted graph, where nodes represent trust dimensions (termed trust aspects), and edges capture their interdependencies. This approach allows smart objects to infer unknown trust dimensions based on available information from other related dimensions.
The proposed model, named T-pattern [11], is designed to capture interdependent trust dimensions. It employs a formalism similar to logical rules. For example, to express that the h o n e s t y dimension derives from r e l i a b i l i t y , we use the rule r e l i a b i l i t y z h o n e s t y . Here, dimensions such as r e l i a b i l i t y and h o n e s t y are not mere logical literals but real variables with values ranging from 1 to 5, where 5 (and 1) denote maximum (and minimum) trustworthiness, respectively. The parameter z is a real value quantifying the inferred trust variable.This approach allows multiple trust variables to contribute to the derivation of another, avoiding the complexities of traditional logic programming by adjusting z values. Furthermore, we introduce the T-Pattern Network (TPN), a framework that represents trust and reputation values alongside dependencies among trust dimensions across all smart objects.
In prior work, we presented the theoretical foundation of the T-pattern model [11] but did not demonstrate its capacity for trust inference. To fully implement the T-pattern model and network, inference techniques are required to derive logical rules, potentially utilizing neural networks, fuzzy logic, Bayesian methods, or other approaches.
To validate the T-pattern model’s ability to infer trust dimensions, we developed its first real implementation. For this purpose, we employed CILIOS (Connectionist Learning and Inter-Ontology Similarities) [12] as the inference engine. CILIOS is a connectionist learning approach that generates logical rules to model agent behavior using a concept graph. It leverages neural-symbolic networks where input and output nodes correspond to logical variables. By observing actor behavior, CILIOS autonomously derives ontologies, requiring only the definition of relevant concepts and categories by the model designer. Using this CILIOS-based implementation, we conducted experiments to assess the model’s effectiveness in enabling smart objects to estimate their partners’ trustworthiness.
The obtained results confirm the T-pattern model’s ability to accurately represent dependencies between trust dimensions—an anticipated yet unverified outcome before our experiments. Different inference engines could be integrated into the T-pattern model depending on various application requirements, such as scalability, memory efficiency, or processing time. However, addressing these aspects goes beyond the scope of this work and represents a challenge for future research.

1.1. Advantages of T-Patterns in the IoT Context

As discussed, most existing models represent trust and reputation as independent scalar values aggregated into a trust vector. This assumption simplifies real-world interactions, where trust dimensions (e.g., reliability and honesty) can be interdependent. In highly dynamic IoT environments, where agent populations evolve rapidly and newcomers frequently join, information about certain trust dimensions may be initially unavailable, while other dimensions can be inferred from initial interactions. Our research addresses this issue by introducing the following advantages:
  • The T-pattern model enables IoT systems to infer unknown trust dimensions, such as reliability, based on available information from related dimensions like honesty.
  • It employs an interpretable logical formalism that supports interactions between smart objects at the IoT edge, while the inference process runs in the background at the cloud level.
  • T-patterns can help smart objects in an IoT environment to operate in presence of noisy or incomplete information since the smart object can use T-patterns to derive an unknown or uncertain information from another one available or more reliable. Our experiments, although limited to a simulated environment, demonstrated that T-patterns can represent the logical links by the different trust dimensions with an high degree of precision, due to the observation of the whole agent community; thus, an inexpert agent, a newcomer, or an agent that operates in presence of noise and that has not a complete information on a given trust dimension can be advantaged by the possibility to derive it form the T-patterns.
  • Our work presents the first real implementation of the T-pattern model, utilizing a neural-symbolic inference engine.
  • CILIOS is specifically designed to extract logical rules by directly observing the behavior of a set of agents in a multi-agent system, regardless of the particular network environment and trust conditions, as the logical rules are extracted from a neural network trained at the cloud layer of the multi-agent system and are continuously updated as agents move from one given network environment to another. This feature made CILIOS particularly suitable to be applied to extract T-patterns in an IoT context.

1.2. Plan of the Paper

This paper is organized as follows. In Section 2, we review related work. Section 3 introduces the scenario under study as well as the architecture of our IoT community and in Section 4, we provide an overview of the T-pattern model proposal. In Section 5, we describe the architecture and the neural-symbolic implementation of T-pattern based on CILIOS and present the experiments carried out and their results that demonstrated the effectiveness of this first version of T-pattern to infer trust. Finally, Section 6 concludes this paper and discusses directions for our ongoing research.

3. The Scenario

In this Section (for a complete list of symbols used in this paper, please refer to Appendix A), we present the considered environment (E) populated by smart IoT objects and users. In such a context, let o represent a smart IoT object and O denote the set of smart IoT objects living within E. Indeed, in networks with 16,000 smart objects, the average error is approximately 5 % . Each object o O is associated with a user and acts on their behalf as a prosumer, both producing and consuming data through interactions with other devices in O.
To enable smooth interactions within E, smart IoT objects should establish trust in one another concerning one or more trust issues, each representing a distinct aspect of their mutual interactions within E. We define Δ = δ 1 , δ 2 , , δ n as a set of n trust issues associated with O in E. Each trust issue δ is characterized by properties such as Expertise, Honesty, Security, and Reliability, which hold the following meanings for a generic smart IoT object o:
  • Expertise. Refers to the level of competence o has in providing knowledgeable opinions within a specific domain. For example, the o’s ability to advise its user on corporate stocks.
  • Honesty. Indicates o’s commitment to truthful behavior, free from deception or misleading practices.
  • Security. Relates to o’s handling of confidential data, including safeguarding it from unauthorized access.
  • Reliability. Measures the consistency and dependability of the services provided by o, with respect to efficiency and effectiveness.
To enable each smart IoT object to quantitatively assess each trust issue in Δ , we introduce the notion of confidence  ϕ in a trust aspect within E. We define ϕ with respect to a group of smart objects in E, denoted as Λ , which, in some cases, may encompass the entire set O E of smart IoT objects (i.e., Λ O ). Specifically, the confidence that a smart IoT object o 1 assigns to a trust issue δ i Δ of another smart IoT object o 2 within the group Λ is represented by ϕ o 1 , o 2 , Λ ( δ i ) . This confidence can take either (i) a real value within [ 1 , 5 ] , where 1 (resp., 5) represents the minimum (resp., maximum) trust level of ϕ o 1 , o 2 , Λ ( δ i ) , or (ii) a value of n u l l if ϕ has not yet been assessed.
Moreover, within a group Λ O , the confidence perceived by the members of Λ concerning a trust issue δ i for a smart IoT object o 1 (denoted ϕ o 1 , o 1 , Λ ( δ i ) ) may serve as the group reputation of Λ with respect to δ i . Like individual confidence, this value can range within [ 1 , 5 ] or remain n u l l if it has not yet been evaluated. It is worth noting that if Λ coincides with O, then ϕ o 1 , o 1 , Λ ( δ i ) will represent the community reputation of the all smart IoT objects regarding the trust issue δ i .

4. A T-Pattern Model Overview

In this section, we provide an overview of the T-pattern model [11].
Before introducing the mathematical formalism, in order to make it clearer and more understandable, we provide a leading example involving two smart IoT objects, denoted as o 1 and o 2 (Figure 1) within an environment E represented by the tuple O , Δ , Φ , τ , where
Figure 1. An example of a simple TPN with two smart IoT objects (i.e., o 1 and o 2 ) considering expertise (X), honesty (H), reliability (R), and security (S) trust issues.
  • O = { o 1 , o 2 } is a set of smart IoT objects;
  • G = is a set of groups (in this example, no specific groups are considered, so G = { } );
  • Δ = { R , H , S , X } is a set of trust issues, namely reliability (R), honesty (H), security (S), and expertise (X);
  • Φ maps the mutual confidence values among the smart IoT objects. In this example, each confidence value is set to null, assuming there is no initial knowledge about the mutual trustworthiness of the other smart IoT objects. (For simplicity, we assume the existence of a single group representing the entire social community E, and thus Λ is omitted).
  • τ is a set consisting of two T-patterns (i.e., the links o 1 , o 2 , N o 1 , o 2 and o 2 , o 1 , N o 2 , o 1 in Figure 1, which represent the initial unsymmetrical knowledge of how the two smart IoT objects perceive each other’s trustworthiness).
In the proposed example, network N o 1 , o 2 (on the left in Figure 1) shows that o 1 derives the honesty of o 2 from its reliability using the derivation rule R 0.7 H , with a ratio of 0.7 . In other words, o 1 trusts o 2 ’s honesty at 70 % of its reliability level, even without prior direct experience of o 2 ’s honesty, provided that o 2 ’s reliability has been verified. Similarly, N o 1 , o 2 illustrates that o 1 infers the security of o 2 based on its honesty using the derivation rule H 0.9 S , with a ratio of 0.9 , again without any previous experience regarding o 2 ’s security. Finally, in the right part of Figure 1, the network N o 2 , o 1 reflects the following derivation rules, with the respective ratios indicated.
R 0.7 H ; H 0.9 S ; H 0.6 X ; X 0.8 R
In our framework, a P2P T-pattern is formally defined as a tuple τ = N Δ , o 1 , o 2 , Λ , where N Δ represents a network that describes peer-to-peer trust relationships for the ordered pair of smart IoT objects o 1 and o 2 , in the context of both the group Λ of smart IoT objects and the group Δ of trust issues. In essence, a P2P T-pattern captures how o 1 perceives trust in o 2 within the context of a group Λ for specific trust issues.
Formally, the network N Δ = V , L consists of a set of trust issues V Δ and a set L of links between these trust issues. For any pair of trust issues δ 1 , δ 2 Δ , the link l L is represented by the ordered tuple l = δ 1 , δ 2 , Λ , w , where Λ is the group context and w is a real-valued weight in the range [ 0 , 1 ] . In the context of a P2P T-pattern, the weight w quantifies the perceived strength of the relationship between trust issues δ 1 and δ 2 as assessed by o 1 in regard to o 2 , and is calculated as follows:
w = ϕ o 1 , o 2 , Λ ( δ 1 ) ϕ o 1 , o 2 , Λ ( δ 2 )
which implies that ϕ o 1 , o 2 , Λ ( δ 1 ) = w · ϕ o 1 , o 2 , Λ ( δ 2 ) . For example, if the directed link between trust issues δ 1 and δ 2 has an associated weight w = 0.65 , this indicates that the confidence in δ 2 is 65% of the confidence assigned to δ 1 .
We also define a global T-pattern as a tuple τ = N Δ , o 1 , o 1 , Λ , representing a single smart IoT object evaluated by the entire group Λ , where Λ O . The weight w, in this case, reflects the collective perception of the relationship between trust issues δ 1 and δ 2 regarding smart object o, and it is computed similarly to Equation (2): ϕ o , o , Λ ( δ 1 ) = w · ϕ o , o , Λ ( δ 2 ) .
To manage a P2P T-pattern for a generic link l = δ 1 , δ 2 , Λ , w within the network N Δ , we apply the following three rules:
  • Derivation rule (DR). It computes ϕ o 1 , o 2 , Λ ( δ 1 ) as w · ϕ o 1 , o 2 , Λ ( δ 2 ) . It is denoted as follows:
    δ 1 w δ 2
  • v-Assignment rule (VR). It assigns a value v [ 1 , 5 ] { n u l l } to the confidence γ δ 1 , δ 2 , Λ ( o ) . It is denoted as follows:
    v o
  • w-Assignment (ZR). It joins a value c [ 0 , 1 ] to the value w of a link l. It is denoted as follows:
    c w
These three rules can be automatically applied to both P2P and global T-pattern:
  • When, due to Equation (3) or other updates, the confidence ϕ o 1 , o 2 , Λ ( δ ) (or, respectively, ϕ o , o , Λ ( δ ) ) for a trust issue δ is updated, then Equation (2) is applied to each link originating from δ represented in N Δ and is propagated to the other nodes linked to the updated nodes, excluding δ .
  • For each link l = δ , δ * , Λ , w directed towards δ * , Equation (5) is automatically applied, updating w to c, which is calculated as the ratio ϕ o 1 , o 2 , Λ ( δ ) ϕ o 1 , o 2 , Λ ( δ * ) .
  • Each time the weight w of a link l is updated by Equation (5), then Equation (2) is automatically applied to the link’s destination node and is propagated to other nodes linked to the updated nodes, excluding the origin node of link l.
This procedure can be synthetically described by the pseudo-code shown in Algorithm 1.
Algorithm 1 Trust propagation algorithm
  • for all trust issue δ do
  •   if  ϕ o 1 , o 2 , Λ ( δ ) OR ϕ o , o , Λ ( δ ) is updated
  •     for all link l originating from δNΔ do
  •       Apply and propagate Equation (2)
  •     end for
  •     for all link l = δ , δ * , Λ , w directed towards δ *  do
  •       Apply Equation (5) automatically
  •     end for
  •     for all link l do
  •       if weight w of l is updated by Equation (5) then
  •        Apply and propagate Equation (2) automatically
  •       end if
  •     end for
  •   end if
  • end for
In such a way, the T-pattern model can be applied to a large variety of scenarios, from more simple to more complex or adversarial ones (e.g., untrustworthy devices or sudden changes in behavior).
Finally, we define a T-Pattern Network (TPN) as a tuple O , G , Φ , Λ , T , where O is the set of smart IoT objects, G represents groups (i.e., G = Λ 1 , Λ 2 , , Λ n ), Φ maps each confidence ϕ o 1 , o 2 , Λ to a value in [ 1 , 5 ] n u l l , Λ is the set of trust issues, and T is the set of T-pattern on Δ , ensuring that no two T-patterns in T are associated with the same triplet ( o 1 , o 2 , Λ ) . A T-pattern can be considered a link between two smart IoT objects; therefore, a TPN can be viewed as a network where the vertices are smart IoT objects, and the links represent the T-pattern in T. Each link (i.e., T-pattern) has an associated weight defined by ( Φ , G Δ ) . Based on the above automatic rule activation, the consistency of all T-pattern with the confidence mapping ( Γ ) is maintained. We highlight that all the trust measures are dimensionless, having values ranging in the interval [1…5], where 1 (resp. 5) means minimum (resp. maximum) trust, and for this reason, in the introduced formulas, there is not ever a division by zero.

5. A Neural-Symbolic Implementation of T-Pattern

As discussed in the introductory section, while in [11], we only presented the T-pattern model, in this work, we introduce its first implementation based on the T-Pattern Architecture, equipped with the neural-symbolic inference engine CILIOS [12]. This implementation allowed us to conduct an experimental campaign in a simulated environment populated with smart IoT objects. The goal of such experiments was to verify the effectiveness of the T-pattern model in inferring trust.
Therefore, this section provides a detailed description of the experiments conducted and their corresponding results.

5.1. The T-Pattern Architecture

The T-Pattern Architecture (TPA) is a multi-SO architecture designed to manage a T-Pattern Network N E T = O , G , Δ , Γ , P , deriving the information needed to update T-patterns by observing the behavior of smart IoT objects and where O is the set of smart objects, G is a set of groups ( Ω 1 , Ω 2 , , Ω n ) , D e l t a is the set of trust issues, Γ is a mapping on a confidence γ o 1 , o 2 , Λ which gives a value ranging in [ 1 , 5 ] , and P is a set of T-patterns on Δ , such that there are not two T-patterns belonging to P associated with the same ordered triplet ( o 1 , o 2 , Λ ) .
The TPA is distributed across three logical levels (see Figure 2), where each level follows the automated rules described in Section 4 (see also [11] for additional details):
Figure 2. The three-layer TPA architecture.
  • A smart object level, consisting of n trust manager smart objects  t m 1 , t m 2 , , t m n , where each t m i is associated with a corresponding smart object o i O of N E T and is capable of updating the trust patterns associated with all edges originating from s i in N E T ;
  • A group level, consisting of l group manager smart objects  g m 1 , g m 2 , , g m l , where each g m i is associated with the corresponding group Ω i G of N E T and is capable of calculating the group reputation γ o , o , Ω i ( t ) for each smart object o Ω i and each trust aspect k K ;
  • A community level, consisting of a single community manager  C M of smart objects, which is capable of calculating the community reputation γ o , o , S ( t ) for each smart object o O and each trust aspect δ Δ .
The three layers can be managed at three different computational layers: (i) the smart object level is represented by the smart object (IoT), and (ii) the group level can be managed into the Fog [50] and the community level can be mapped into the Cloud.
We highlight that the presence of the three layers gives the possibility of managing each layer with different and independent computational resources as follows: (i) the smart object level is managed by the smart objects, (ii) the group level is managed by the computational power of the Fog, and, finally, the community level is managed into the Cloud. This way, increasing the number of smart objects or object groups do not significantly impact system performance. In particular, the scalability of the upper service will be easily ensured by the elasticity of Cloud resources.

5.2. The Inference Engine

The T-pattern model requires an inference technique in order to derive suitable logical rules. To this end, we identified in CILIOS [12] a promising inference engine for T-pattern. A brief description of CILIOS is provided below.
CILIOS (Connectionist Inductive Learning and Inter-Ontology Similarities) is a system originally developed to enhance agent collaboration in multi-agent environments (MAS). This system enables the construction of knowledge representations, known as ontologies, that support collaborative decision-making and personalized recommendations.
The architecture of CILIOS combines connectionist learning techniques with symbolic ontology models. In CILIOS, observed behaviors are translated into logical rules using neural-symbolic networks. Each CILIOS agent can dynamically represent and update its own ontology, which describes the relevant concepts and categories for interaction.
By integrating ontologies, CILIOS facilitates efficient cooperation, adapts to changes, and continuously improves interaction capabilities, increasing the system’s ability to respond to complex contexts like the Web and IoT networks. Furthermore, the system is designed to be flexible and scalable, supporting a broad range of applications where continuous learning and dynamic adaptation are essential.
In order to understand how CILIOS can be used to extract T-patterns (a complete description can be found in [12]), we briefly describe the underlying idea on which it is conceived. This idea is based on the possibility that symbolic knowledge can be represented by a connectionist system, as a neural network, in order to build an effective learning system. In particular, it is proved that, for each extended logic program P, there exists a feed-forward neural network N with exactly one hidden layer and semi-linear activation functions, which is equivalent to P in the sense that N computes the model of P. In our current application to the IoT described in this paper, a set of T-patterns as, for example, that represented in Equation (2) can be viewed as a logic program P, and it can be derived following the constructive definition provided in [12]. In particular, this construction passes through the construction of a feed-forward neural network that will be trained on the data provided by the smart objects’ interactions and that will be then analyzed to be transformed in a logic program representing the T-patterns.

5.3. The Simulation

To evaluate the effectiveness of the T-pattern model in representing trust relationships in a simulated smart IoT environment, we considered the following elements:
  • A set of n smart IoT objects O;
  • A set of k groups G;
  • A set of trust issues Δ = { R , H , S , X } , representing reliability (R), honesty (H), security (S), and expertise (X).

5.4. Training Phase

We simulated i mutual interactions between the smart IoT objects in O, applying the CILIOS approach to derive the set τ (containing the extracted T-pattern), as described in Section 4. The set Φ (mapping mutual confidence values among smart IoT objects) was initially randomly generated and refined continuously throughout the training process. We repeated the training phase for different values of n = 2000 , 4000 , 8000 , 16,000 and k = 5 , 10 , 15 , 20 , resulting in 16 training scenarios, each denoted by T R A I N I N G n , k . We used i = 100,000 (resp. 200,000, 400,000, 800,000) for n = 2000 (resp. 4000, 8000, 16,000), assuming that interactions increase linearly with the number of smart IoT objects.

5.5. Test Phase

For each training phase T R A I N I N G n , k , we conducted a corresponding test phase T E S T n , k , simulating i mutual interactions among the smart IoT objects in O. For each pair of smart IoT objects o h , o k , only the reliability value was initially known, while honesty, security, and expertise were derived using the T-pattern extracted in the training phase.
As in the training phase, we used i = 100,000 (resp. 200,000, 400,000, 800,000) for n = 2000 (resp. 4000, 8000, 16,000). For each test phase T E S T n , k and each pair of smart IoT objects o p , o q , we calculated the percentage error e ( o p , o q , H O N ) (resp. e ( o p , o q , S E C ) , e ( o p , o q , E X P ) ) by comparing the inferred values for honesty, security, and expertise to the actual values. Finally, we computed the average percentage error e ( H O N ) (resp. e ( S E C ) , e ( E X P ) ) across all pairs o p , o q , with the results presented in Table 2, Table 3 and Table 4.
Table 2. Percentage error e ( H O N ) for different values of number of smart IoT objects n and number of groups k.
Table 3. Percentage error e ( S E C ) for different values of number of smart IoT objects n and number of groups k.
Table 4. Percentage error e ( E X P ) for different values of number of smart IoT objects n and number of groups k.
The same results are graphically represented in the bar diagrams in Figure 3, Figure 4 and Figure 5.
Figure 3. Percentage error e ( H O N ) for different values of number of smart IoT objects n and number of groups k.
Figure 4. Percentage error e ( S E C ) for different values of number of smart IoT objects n and number of groups k.
Figure 5. Percentage error e ( E X P ) for different values of number of smart IoT objects n and number of groups k.

5.6. Discussion of the Results

The experimental results show that the T-pattern model enables smart IoT objects to estimate confidence values in their partners for three trust issues (honesty, security, and expertise) derived from reliability values, with an average error ranging from 5 % to 15 % with respect to actual values, demonstrating that the model has a sufficient accuracy. Accuracy improves with network size: in networks with 16,000 smart objects, the average error is approximately 5 % . In particular, in our simulations, larger networks collected more interactions (the number i), thus providing a more precise inference mechanism during training with respect to small networks. This results do not represent a limitation for small networks because, as we verified in our simulations, even smaller networks, sooner or later, will reach a number of interactions to decrease the measured error.
We also observe that the average error increases with the number of groups, as a higher number of groups complicates the inference process. While the average error with only 5 groups ranges from 5 % to 11 % , it increases to between 11 % and 15 % with 20 groups. We highlight that increasing the number of groups imply increasing the number of logical rules that the inference engine has to extract. It has been demonstrated in [12] that the effectiveness of the logical rules in representing the data with which the neural network is trained decreases with the number of logical rules; thus, this result was expected.
As we further discuss in our conclusion, we aim to improve this particular aspect in the near future. We finally observe that the experimental results are consistent across all trust issues derived using the T-pattern model.

5.7. Efficiency and Costs

We highlight that the construction of T-patterns is performed at the cloud level, in background with respect to the ordinary activities of the IoT systems, and thus, it does not significantly impact the execution time of the other tasks involving the smart objects, while T-patterns model consumes a minimal amount of system memory, relating to the storage of a graph of relatively small dimension. This implies that no significant barriers exist for the implementation of this approach to large-scale wireless IoT systems. In order to compute T-patterns using the CILIOS inference engine, we have used, in pur simulation at the cloud level, a 64-bit workstation with a Intel(R) Core(TM) i7-10875H CPU @ 2.30 GHz and 128 GB of RAM. Table 5 shows how computation time changes with varying network sizes.
Table 5. Execution time t (in minutes) of the T-pattern model construction for different values of number of smart IoT objects n and number of groups k.
As a final consideration, we highlight that a different group size impacts the system performance simply based on its impact on the global size n of the multi-agent community. This is derived from several simulations that we performed by fixing k and increasing the size of a given group (that results in an increment of the parameter n). We simply re-obtained the results shown in the tables above.

6. Conclusions

An important emerging aspect in IoT smart object communities is the representation of mutual trust among community members. To address the limitations of current approaches in the literature, which often struggle to capture complex trust relationships in IoT smart object communities, we previously developed a model called the T-pattern. This model maintains trust information through a weighted graph, where nodes represent trust dimensions and edges represent relationships between these dimensions. The strength of the T-pattern model lies in its ability to derive one unknown trust dimension from another by leveraging their interdependence. To validate the effectiveness of T-patterns in inferring trust dimensions, this paper presents experiments conducted with an implementation of this model that uses a neural-symbolic approach as inference engine, which is based on CILIOS. These experiments, designed to estimate trust values for three specific trust aspects—honesty, safety, and competence—based on trustworthiness values, demonstrate the effectiveness and accuracy of this T-pattern implementation in reliably inferring trust.
We note that the calculation of trust measures is the hard part of the presented approach and that the use of neural networks is performed offline. However, we have shown how the training of the model can be executed in the background, while the agents interact in the IoT scenario, using the cloud level of the IoT system without interfering with the ordinary agent processes. The trained model is used when the training is completed and the simulation we have performed show that its exploitation is useful to improve the capability of the system to accurately determine the trust of the different smart objects.
Building on this line of research, after having verified the potentiality of our T-pattern model, our future studies will focus on implementing the proposed model by adopting new inference engines and considering additional scenarios. Moreover, for a more exhaustive analysis, we aim to evaluate our model under different error metrics and assessing its effectiveness with respect to other issues like computational efficiency (e.g., inference time and memory usage), robustness under noise or malicious input, and sensitivity to parameter changes.
Finally, we highlight that our proposal to apply T-patterns to the IoT environment is the first attempt we make at applying our trust model to a complex and distributed Information Systems scenario, and the simulations we have performed should be considered only as a verification of the feasibility of the idea. In our ongoing research, we are planning to apply the T-pattern model to a real IoT environment, directly observing the behavior of smart objects when interacting with each other, constructing the T-patterns based on these observations, and then computing the effectiveness of the model. Moving from simulations to a real environment implies facing novel issues that we did not deal with in this paper, as the management of real-word data noise and the scalability of the system with respect to the dimensions of the smart object population. Secondly, we highlight that the architecture of our system, which is delegated to the cloud-level construction of T-patterns with an inference engine that is used in the background with respect to other ordinary IoT activities, possess the capability to integrate with edge or fog computing frameworks as a system for performing real-time trust estimation. Moreover, it is also possible to study the possibility of combining T-pattern with reinforcement learning integration to create hybrid trust models that might enhance inference accuracy levels.

Author Contributions

All authors collaborated equally on this article. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Italian Ministry of University and Research (MUR) Project “T-LADIES” under Grant PRIN 2020TL3X8X and in part by Pia.ce.ri. 2024–2026 funded by the University of Catania and in part by the Project CAL.HUB.RIA funded by the Italian Ministry of Health, Project CUP: F63C22000530001. Local Project CUP: C33C22000540001.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Symbol List

For the sake of clarity, the list of adopted symbols is provided below.
Table A1. Symbol list.
Table A1. Symbol list.
SymbolMeaning
Ethe Environment
oa smart IoT object
Oa set of smart IoT objects
δ a trust issue
Δ a set of trust issues
Λ a group of smart IoT objects
Ga set of groups of smart IoT objects
ϕ a confidence value
Φ a set of confidence values
ϕ o 1 , o 2 , Λ ( δ i ) the confidence of o 1 for δ i w.r.t. o 2 within the group Λ
ϕ o 1 , o 1 , Λ ( δ i ) the confidence of the group Λ for δ i w.r.t. o 1
Γ a mapping on a confidence γ o 1 , o 2 , Λ which gives a value ranging in [ 1 , 5 ]
τ a T-pattern τ = o 1 , o 2 , Λ , N Δ over two smart objects o 1 and o 2 , a trust network N Δ within a group Λ
Ta set of T-pattern
N Δ a weighted directed graph representing trust relationships
va vertex in N Δ
Va set of vertexes in N Δ
la link in N Δ
La set of links in N Δ
wa weight on a link l N Δ

References

  1. Telang, P.; Singh, M.P.; Yorke-Smith, N. Maintenance of Social Commitments in Multiagent Systems. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 11369–11377. [Google Scholar]
  2. Jaques, N.; Lazaridou, A.; Hughes, E.; Gulcehre, C.; Ortega, P.; Strouse, D.; Leibo, J.Z.; De Freitas, N. Social influence as intrinsic motivation for multi-agent deep reinforcement learning. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 3040–3049. [Google Scholar]
  3. Esmaeili, A.; Mozayani, N.; Motlagh, M.R.J.; Matson, E.T. A socially-based distributed self-organizing algorithm for holonic multi-agent systems: Case study in a task environment. Cogn. Syst. Res. 2017, 43, 21–44. [Google Scholar] [CrossRef]
  4. Walczak, S. Society of Agents: A framework for multi-agent collaborative problem solving. In Natural Language Processing: Concepts, Methodologies, Tools, and Applications; IGI Global: Hershey, PA, USA, 2020; pp. 160–183. [Google Scholar]
  5. Torreño, A.; Onaindia, E.; Komenda, A.; Štolba, M. Cooperative multi-agent planning: A survey. ACM Comput. Surv. (CSUR) 2017, 50, 1–32. [Google Scholar] [CrossRef]
  6. Khan, W.Z.; Aalsalem, M.Y.; Khan, M.K.; Arshad, Q. When social objects collaborate: Concepts, processing elements, attacks and challenges. Comput. Electr. Eng. 2017, 58, 397–411. [Google Scholar] [CrossRef]
  7. Jafari, S.; Navidi, H. A game-theoretic approach for modeling competitive diffusion over social networks. Games 2018, 9, 8. [Google Scholar] [CrossRef]
  8. He, Z.; Han, G.; Cheng, T.; Fan, B.; Dong, J. Evolutionary food quality and location strategies for restaurants in competitive online-to-offline food ordering and delivery markets: An agent-based approach. Int. J. Prod. Econ. 2019, 215, 61–72. [Google Scholar] [CrossRef]
  9. Kowshalya, A.M.; Valarmathi, M. Trust management for reliable decision making among social objects in the Social Internet of Things. IET Netw. 2017, 6, 75–80. [Google Scholar] [CrossRef]
  10. Brogan, C.; Smith, J. Trust Agents: Using the Web to Build Influence, Improve Reputation, and Earn Trust; John Wiley & Sons: New York, NY, USA, 2020. [Google Scholar]
  11. Messina, F.; Rosaci, D.; Sarnè, G.M. Applying Trust Patterns to Model Complex Trustworthiness in the Internet of Things. Electronics 2024, 13, 2107. [Google Scholar] [CrossRef]
  12. Rosaci, D. CILIOS: Connectionist inductive learning and inter-ontology similarities for recommending information agents. Inf. Syst. 2007, 32, 793–825. [Google Scholar] [CrossRef]
  13. Demolombe, R. Reasoning about trust: A formal logical framework. In Proceedings of the International Conference on Trust Management; Springer: Berlin/Heidelberg, Germany, 2004; pp. 291–303. [Google Scholar]
  14. Drawel, N.; Bentahar, J.; Qu, H. Computationally Grounded Quantitative Trust with Time. In Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems, Auckland, New Zealand, 9–13 May 2020; pp. 1837–1839. [Google Scholar]
  15. Baier, C.; Katoen, J.P. Principles of Model Checking; MIT Press: Cambridge, MA, USA, 2008. [Google Scholar]
  16. Liu, X.; Datta, A.; Lim, E.P. Computational Trust Models and Machine Learning; CRC Press: Boca Raton, FL, USA, 2014. [Google Scholar]
  17. Ma, W.; Wang, X.; Hu, M.; Zhou, Q. Machine learning empowered trust evaluation method for IoT devices. IEEE Access 2021, 9, 65066–65077. [Google Scholar] [CrossRef]
  18. Wang, J.; Jing, X.; Yan, Z.; Fu, Y.; Pedrycz, W.; Yang, L.T. A survey on trust evaluation based on machine learning. ACM Comput. Surv. (CSUR) 2020, 53, 1–36. [Google Scholar] [CrossRef]
  19. Palmer-Brown, D. Neural Networks for Modal and Virtual Learning. In Proceedings of the Artificial Intelligence Applications and Innovations III; Iliadis, L., Vlahavas, I., Bramer, M., Eds.; Springer: Boston, MA, USA, 2009; p. 2. [Google Scholar]
  20. Liu, G.; Li, C.; Yang, Q. Neuralwalk: Trust assessment in online social networks with neural networks. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; pp. 1999–2007. [Google Scholar]
  21. Sagar, S.; Mahmood, A.; Wang, K.; Sheng, Q.Z.; Pabani, J.K.; Zhang, W.E. Trust–SIoT: Toward trustworthy object classification in the social internet of things. IEEE Trans. Netw. Serv. Manag. 2023, 20, 1210–1223. [Google Scholar] [CrossRef]
  22. Awan, K.A.; Din, I.U.; Almogren, A.; Almajed, H.; Mohiuddin, I.; Guizani, M. NeuroTrust—Artificial-neural-network-based intelligent trust management mechanism for large-scale Internet of Medical Things. IEEE Internet Things J. 2020, 8, 15672–15682. [Google Scholar] [CrossRef]
  23. Bhor, H.N.; Kalla, M. TRUST-based features for detecting the intruders in the Internet of Things network using deep learning. Comput. Intell. 2022, 38, 438–462. [Google Scholar] [CrossRef]
  24. Sharma, A.; Pilli, E.S.; Mazumdar, A.P.; Gera, P. Towards trustworthy Internet of Things: A survey on Trust Management applications and schemes. Comput. Commun. 2020, 160, 475–493. [Google Scholar] [CrossRef]
  25. Hussain, Y.; Zhiqiu, H.; Akbar, M.A.; Alsanad, A.; Alsanad, A.A.A.; Nawaz, A.; Khan, I.A.; Khan, Z.U. Context-aware trust and reputation model for fog-based IoT. IEEE Access 2020, 8, 31622–31632. [Google Scholar] [CrossRef]
  26. Fortino, G.; Fotia, L.; Messina, F.; Rosaci, D.; Sarné, G.M.L. Trust and reputation in the internet of things: State-of-the-art and research challenges. IEEE Access 2020, 8, 60117–60125. [Google Scholar] [CrossRef]
  27. Fortino, G.; Fotia, L.; Messina, F.; Rosaci, D.; Sarné, G.M.L. A meritocratic trust-based group formation in an IoT environment for smart cities. Future Gener. Comput. Syst. 2020, 108, 34–45. [Google Scholar] [CrossRef]
  28. Guo, J.; Liu, Z.; Tian, S.; Huang, F.; Li, J.; Li, X.; Igorevich, K.K.; Ma, J. TFL-DT: A Trust Evaluation Scheme for Federated Learning in Digital Twin for Mobile Networks. IEEE J. Sel. Areas Commun. 2023, 41, 3548–3560. [Google Scholar] [CrossRef]
  29. Guarino, N. Formal Ontology in Information Systems: Proceedings of the First International Conference (FOIS’98), June 6–8, Trento, Italy; IOS Press: Amsterdam, The Netherlands, 1998; Volume 46. [Google Scholar]
  30. Java Agent DEvelopment Framework (JADE). 2024. Available online: http://jade.tilab.com/ (accessed on 10 January 2025).
  31. Viljanen, L. Towards an ontology of trust. In Proceedings of the International Conference on Trust, Privacy and Security in Digital Business; Springer: Berlin/Heidelberg, Germany, 2005; pp. 175–184. [Google Scholar]
  32. Nagy, M.; Vargas-Vera, M.; Motta, E. Multi-agent Conflict Resolution with Trust for Ontology Mapping. In Proceedings of the Intelligent Distributed Computing, Systems and Applications: Proceedings of the 2nd International Symposium on Intelligent Distributed Computing–IDC 2008, Catania, Italy, 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 275–280. [Google Scholar]
  33. Sadigh, B.L.; Unver, H.O.; Nikghadam, S.; Dogdu, E.; Ozbayoglu, A.M.; Kilic, S.E. An ontology-based multi-agent virtual enterprise system (OMAVE): Part 1: Domain model l ing and rule management. Int. J. Comput. Integr. Manuf. 2017, 30, 320–343. [Google Scholar] [CrossRef]
  34. Amaral, G.; Sales, T.P.; Guizzardi, G.; Porello, D. Towards a reference ontology of trust. In Proceedings of the On the Move to Meaningful Internet Systems: OTM 2019 Conferences: Confederated International Conferences: CoopIS, ODBASE, C&TC 2019, Rhodes, Greece, 21–25 October 2019; Proceedings. Springer: Berlin/Heidelberg, Germany, 2019; pp. 3–21. [Google Scholar]
  35. Barbosa, R.; Santos, R.; Novais, P. Trust-based negotiation in multiagent systems: A systematic review. In Proceedings of the International Conference on Practical Applications of Agents and Multi-Agent Systems; Springer: Berlin/Heidelberg, Germany, 2023; pp. 133–144. [Google Scholar]
  36. Sobh, T.S. A secure and integrated ontology-based fusion using multi-agent system. Int. J. Inf. Commun. Technol. 2024, 25, 48–73. [Google Scholar] [CrossRef]
  37. Zhang, S.; Dong, Y.; Zhang, Y.; Payne, T.R.; Zhang, J. Large Language Model Assisted Multi-Agent Dialogue for Ontology Alignment. In Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems, Auckland, New Zealand, 6–10 May 2024; pp. 2594–2596. [Google Scholar]
  38. Burgess, M. Thinking in Promises: Designing Systems for Cooperation; O’Reilly Media, Inc.: Newton, MA, USA, 2015. [Google Scholar]
  39. Burgess, M.; Fagernes, S. Autonomic pervasive computing: A smart mall scenario using promise theory. In Proceedings of the 1st IEEE International Workshop on Modelling Autonomic Communications Environments (MACE), Dublin, Ireland, 25–26 October 2006; pp. 133–160. [Google Scholar]
  40. Burgess, M.; Dunbar, R.I. A promise theory perspective on the role of intent in group dynamics. arXiv 2024, arXiv:2402.00598. [Google Scholar]
  41. Burgess, M.; Dunbar, R.I. Group related phenomena in wikipedia edits. arXiv 2024, arXiv:2402.00595. [Google Scholar]
  42. Firesmith, D.; Henderson-Sellers, B.; Graham, I. OPEN Modeling Language (OML) Reference Manual; CUP Archive: Cambridge, UK, 1998. [Google Scholar]
  43. McGuinness, D.L.; Fikes, R.; Hendler, J.; Stein, L.A. DAML+ OIL: An ontology language for the Semantic Web. IEEE Intell. Syst. 2002, 17, 72–80. [Google Scholar] [CrossRef]
  44. Horrocks, I.; Patel-Schneider, P.F.; Boley, H.; Tabet, S.; Grosof, B.; Dean, M. SWRL: A semantic web rule language combining OWL and RuleML. W3C Memb. Submiss. 2004, 21, 1–31. [Google Scholar]
  45. van der Hoek, W. Logical foundations of agent-based computing. In ECCAI Advanced Course on Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2001; pp. 50–73. [Google Scholar]
  46. Rao, A.S.; Georgeff, M.P. Decision procedures for BDI logics. J. Log. Comput. 1998, 8, 293–343. [Google Scholar] [CrossRef]
  47. Alferes, J.J.; Pereira, L.M. Updates plus preferences. In Proceedings of the Logics in Artificial Intelligence: European Workshop, JELIA 2000, Málaga, Spain, 29 September–2 October 2000; Proceedings 7. Springer: Berlin/Heidelberg, Germany, 2000; pp. 345–360. [Google Scholar]
  48. Extensible Markup Language (XML). 2024. Available online: https://www.w3.org/XML/ (accessed on 10 January 2025).
  49. Liu, M.; Teng, F.; Zhang, Z.; Ge, P.; Sun, M.; Deng, R.; Cheng, P.; Chen, J. Enhancing cyber-resiliency of der-based smart grid: A survey. IEEE Trans. Smart Grid 2024, 15, 4998–5030. [Google Scholar] [CrossRef]
  50. Atlam, H.F.; Walters, R.J.; Wills, G.B. Fog computing and the internet of things: A review. Big Data Cogn. Comput. 2018, 2, 10. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.