Next Article in Journal
Correction: Bianca, B.L.; Gheorghe, P.S. Unsupervised Clustering for Hyperspectral Images. Symmetry 2020, 12, 277
Next Article in Special Issue
q-Rung Orthopair Fuzzy Geometric Aggregation Operators Based on Generalized and Group-Generalized Parameters with Application to Water Loss Management
Previous Article in Journal
A Supersymmetry and Quantum Cryptosystem with Path Integral Approach in Biology
Previous Article in Special Issue
Fuzzy Random Chance-Constrained Programming Model for the Vehicle Routing Problem of Hazardous Materials Transportation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Linear Diophantine Fuzzy Soft Rough Sets for the Selection of Sustainable Material Handling Equipment

1
Department of Mathematics, University of the Punjab, Lahore 54590, Pakistan
2
Department of Mathematics, Khawaja Fareed University of Engineering and Information Technology (KFUEIT), Rahim Yar Khan 64200, Pakistan
3
School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China
4
Department of Logistics, University of Defence, 11000 Belgrade, Serbia
5
Department of Mathematics, Huzhou University, Huzhou 313000, China
6
Hunan Provincial Key Laboratory of Mathematical Modeling and Analysis in Engineering, Changsha University of Science & Technology, Changsha 410114, China
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(8), 1215; https://doi.org/10.3390/sym12081215
Submission received: 25 June 2020 / Revised: 21 July 2020 / Accepted: 22 July 2020 / Published: 24 July 2020
(This article belongs to the Special Issue Uncertain Multi-Criteria Optimization Problems)

Abstract

:
The concept of linear Diophantine fuzzy sets (LDFSs) is a new approach for modeling uncertainties in decision analysis. Due to the addition of reference or control parameters with membership and non-membership grades, LDFS is more flexible and reliable than existing concepts of intuitionistic fuzzy sets (IFSs), Pythagorean fuzzy sets (PFSs), and q-rung orthopair fuzzy sets (q-ROFSs). In this paper, the notions of linear Diophantine fuzzy soft rough sets (LDFSRSs) and soft rough linear Diophantine fuzzy sets (SRLDFSs) are proposed as new hybrid models of soft sets, rough sets, and LDFS. The suggested models of LDFSRSs and SRLDFSs are more flexible to discuss fuzziness and roughness in terms of upper and lower approximation operators. Certain operations on LDFSRSs and SRLDFSs have been established to discuss robust multi-criteria decision making (MCDM) for the selection of sustainable material handling equipment. For these objectives, some algorithms are developed for the ranking of feasible alternatives and deriving an optimal decision. Meanwhile, the ideas of the upper reduct, lower reduct, and core set are defined as key factors in the proposed MCDM technique. An application of MCDM is illustrated by a numerical example, and the final ranking in the selection of sustainable material handling equipment is computed by the proposed algorithms. Finally, a comparison analysis is given to justify the feasibility, reliability, and superiority of the proposed models.

1. Introduction

The multi-criteria decision making (MCDM) techniques have been rigorously investigated by many researchers around the real world. Due to uncertain and vague information, the complexity of human’s decision making has grown broadly in the present era. This pursuit gave rise to many resourceful techniques to deal with real-world problems. The methodologies developed for this objective essentially rely on the description of the problem under contemplation. The problem of imperfect, uncertain, and vague information has been focused on by many researchers in the last few decades.
Zadeh (1965) developed the notion of fuzzy sets, fuzzy numbers, and linguistic variables to describe hidden uncertain information in the objects by using membership grades. Researchers found that membership/affiliation grades alone are not enough to express some real-life situations such as: benefit and loss claims, positive results and side effects of drugs, inferiority and superiority, perfection and imperfection, affiliation and non-affiliation, etc. In order to cope with these challenges, Atanassov (1983) proposed the idea of intuitionistic fuzzy sets (IFSs) with the inclusion of satisfaction or membership grade (MG) and dissatisfaction or non-membership grade (NMG). Yager (2014, 2017) extended IFSs to Pythagorean fuzzy sets (PFSs) and q-rung orthopair fuzzy sets (q-ROFSs).
In inadequate information data, the vagueness caused by the indiscernibility can be manipulated by utilizing the rough set techniques. This is an individualistic generalization of crisp set theory and was first originated by Pawlak (1982). This hypothesis acts as a tool for investigating and implementing solutions for various decision making difficulties found in the fields of computer intelligence, image processing, data analysis, medical sciences, and many more. It eliminates vagueness by using upper and lower approximation operators of a collection by assembling the equivalence relation. The above-listed theories do not deal with the parameterizations of the input information set. For this purpose, Molodtsov (1999) proposed soft set theory to deal with the uncertainties in parametric behavior.
Sustainability is the ability to exist constantly. The main components of sustainability are society, the economy, and the environment. The equilibrium of local and global efforts for sustainability is necessary to meet elementary human needs without destroying the environment. The ability to finance all capital projects is essential for the sustainability of the economy. The environmental concerns while retaining sustainable growth for the environment are becoming increasingly relevant to decision analysis around the world. Sustainability is regarded as a task, process, activity, and exercise through which humankind avoids the destruction of natural resources. The selection of sustainable material handling equipment is essential for the development of infrastructure.
Multi-criteria decision making (MCDM) is a branch of operations research that explicitly evaluates multiple conflicting criteria in decision making. The purpose of MCDM is to support decision makers (DMs) facing problems in ranking feasible alternatives/objects. There are different types of criteria for determining weights for the alternatives/objects. The subjective criteria weights depend on the DM and can change if another DM computes them. On the opposite side, there is the idea of objective weights, which are different because they have the capacity to evaluate alternatives. For determining subjective and objective weights, there are different fuzzy and crisp methods like SWARA (step-wise weight assessment ratio analysis), WASPAS (weighted aggregated sum product assessment), ARAS (additive ratio assessment), AHP (analytic hierarchy process), PIPRECIA (pivot pairwise relative criteria importance assessment), and CRITIC (criteria importance through inter-criteria correlation). Some integrated MCDM methods studied by researchers are TOPSIS (technique for the order preference by similarity to ideal solution), VIKOR (vlse kriterijumska optimizacija kompromisno resenje), PROMETHEE (preference ranking organization method for enrichment evaluations), COPRAS (complex proportional assessment), MOORA (multi-objective optimization by ratio analysis), GRA (grey relational analysis), ANP (analytic network process), BWM (best worst method), and aggregation operators.

1.1. Literature Review

Bellman and Zadeh [1] proposed the MCDM technique based on fuzzy sets for the first time in 1970. Akram et al. [2] introduced the m-polar fuzzy soft rough sets and presented their applications in multi-attribute decision making (MADM) difficulties. Ali et al. [3] established certain properties of rough sets, soft sets, and fuzzy soft sets. Chen and Tan [4] established the concept of the score function, which was presented by Tversky and Kahneman [5] earlier. Feng et al. [6,7,8,9] proposed the idea of soft rough sets. Garg [10] investigated Einstein operators and established the Pythagorean operators to solve decision making obstacles. Hashmi et al. [11] invented the hybrid structure of m-polar neutrosophic set (MPNS) as an abstraction of the bipolar neutrosophic set by combining MPFSsand neutrosophic sets. They developed innovative algorithms to deal with the difficulties in medical sciences and for the clustering of information data. Hashmi and Riaz [12] introduced Pythagorean m-polar fuzzy Dombi operators and proposed a novel technique to the censuses process.
Jose and Kuriaskose [13] proposed the MCDM model for intuitionistic fuzzy numbers (IFNs) by using operators. Naeem et al. [14] introduced multi-criteria group decision making (MCGDM) methods based on TOPSIS and VIKOR using Pythagorean fuzzy soft sets. Pawlak and Skowron [15] presented certain extensions of rough sets.
Riaz and Hashmi [16,17] invented the notions of cubic m-polar fuzzy sets and Pythagorean m-polar fuzzy soft rough sets with applications to decision making difficulties. They established novel structures of soft rough Pythagorean m-polar fuzzy sets and Pythagorean m-polar fuzzy soft rough sets with their applications. Riaz et al. [18,19] introduced the soft rough topology including its applications to group decision making. Riaz and Tehrim [20,21,22] originated the notions of the bipolar fuzzy soft topology, cubic bipolar fuzzy sets, and operators. By using diverse algorithms, they solved some new and challenging decision making applications. Roy et al. [23] introduced a rough strength relational decision making trial and evaluation laboratory (DEMATEL) model for analyzing the key success factors of hospital service quality. Sharma et al. [24] introduced a rough set theory application in forecasting models.
Wei et al. [25] proposed aggregation operators based on hesitant triangular fuzzy information to determine MADM obstacles. Zhang et al. [26] established the concept of intuitionistic fuzzy soft rough sets and presented its applications.
Zhao [27] et al. discovered novel algorithms based on generalized intuitionistic fuzzy aggregation operators. Xu and Chen [28] practiced distance and similarity measures on IFSs. Kulak et al. [29] and Karande et al. [30] prepared some techniques for the assortment of material handling equipment using the information axiom and weighted utility additive theory. Zubair et al. [31] presented the optimization of a material handling system. Vashist [32] presented an algorithm for finding the reduct and core of the information dataset. Zhang et al. [33,34,35] discovered different covering based rough sets, fuzzy rough sets, and intuitionistic fuzzy rough sets with their applications to MADM obstacles. Wang and Triantaphyllou [36] identified irregularities in the ranking when evaluating alternatives using certain elimination et choix traduisant la realite (ELECTRE) methods. In order to evaluate green suppliers, Búyúkózkan and Çifçi [37] presented a novel hybrid MCDM approach based on fuzzy DEMATEL, fuzzy ANP, and fuzzy TOPSIS.
Govindan et al. [38] developed a DEMATEL approach focused on experience to establish sustainability strategies and efficiency in a green supply chain. Via flipped e-learning, Jeong and González-Gómez [39] built a system adjusting to the pedagogical changes in sustainable mathematics education through pre-service teachers (PSTs): rating the requirements with MCDA/F -DEMATEL. Under a q-rung orthopair fuzzy set, Wang and Li [40] established a novel approach for green supplier selection. Xu et al. [41] presented some q-rung dual hesitant Heronian mean operators with their application to multiple group decision making attributes. Soft rough fuzzy sets were developed by Sun and Ma [42] with their applications in strategic decision making. Including its various results and illustrations, Meng et al. [43] introduced the structures of soft rough fuzzy sets and soft fuzzy rough sets. Hussain et al. [44] invented Pythagorean fuzzy soft-rough set models and presented their applications in decision making. Zadeh [45] introduced the concept of a linguistic variable and its application to approximate reasoning.

1.2. Motivation and Objectives

A q-ROFS is the generalization of both IFS and PFS. The main feature of q-ROFS is that the uncertain space for MG and NMG is boarder. Each IFS is a PFS, and each PFS is a q-ROFS, but not conversely. A q-ROFS is more powerful in growing the freedom between MG and NMG. However, there are some situations when these theories are unable to deal with uncertain information. In order to relax existing constraints on MG and NMG, Riaz and Hashmi (2019) introduced the innovative idea of linear Diophantine fuzzy sets (LDFSs). The use of reference or control parameters in LDFS give freedom to DMs in choosing MG and NMG. Moreover, IFSs, PFSs, and q-ROFSs can be considered as specific cases of LDFSs with some limitations (see Figure 1). The semantic comparison of suggested technique with some existing structures is given in Table 1.
The goal of this paper is to develop strong models for MCDM that have less limitations than other models. Table 1 shows the advantages and drawbacks of some set theoretical models. The notions of linear Diophantine fuzzy soft rough sets (LDFSRSs) and soft rough linear Diophantine fuzzy sets (SRLDFSs) are established as new hybrid models of soft sets, rough sets, and LDFSs. The suggested models of LDFSRSs and SRLDFSs are more flexible to discuss fuzziness and roughness in terms of upper and lower approximation operators. Certain operations on LDFSRSs and SRLDFSs have been established to discuss a robust multi-criteria decision making (MCDM) for the selection of sustainable material handling equipment. We present four new algorithms based on LDFS, crisp soft approximation spaces, core sets, and reducts.
The organization of this article is provided as follows. Section 2 implies certain fundamental notions of fuzzy sets, IFSs, PFSs, q-ROFSs, and LDFSs. We investigate fascinating operations and score functions of LDFSs. In Section 3, we invent the notions of LDFSRSs and SRLDFSs by applying the LDFS approximation space and crisp soft approximation space. We establish multiple results based on intended structures with the help of illustrations. In Section 4, we present four novel algorithms to determine the material handling equipment selection obstacle. These algorithms are based on the approximation spaces, score functions, upper and lower reducts, and core set. We examine and compare our suggested structures and their results with certain existing notions. Section 5 provides the conclusion of this manuscript.

2. Some Basic Concepts

First, we assemble fascinating fundamental ideas of LDFSs, rough sets, soft sets, and soft rough sets.
Definition 1
([54]). A linear Diophantine fuzzy set D in Q ¨ is defined as:
D = G , T ¨ D ( G ) , S ¨ D ( G ) , α D ( G ) , β D ( G ) : G Q ¨ ,
where T ¨ D ( G ) , S ¨ D ( G ) , α D ( G ) , β D ( G ) [ 0 , 1 ] are the satisfaction grade, the dissatisfaction grade, and the corresponding reference parameters, respectively. Moreover, it is required that:
0 α D ( G ) + β D ( G ) 1 ,
and:
0 α D ( G ) T ¨ D ( G ) + β D ( G ) S ¨ D ( G ) 1
for all G Q ¨ . The LDFS:
D Q ¨ = { ( G , 1 , 0 , 1 , 0 ) : G Q ¨ }
is called the absolute LDFS in Q ¨ . The LDFS:
D ϕ = { ( G , 0 , 1 , 0 , 1 ) : G Q ¨ }
is called the null LDFS in Q ¨ .
The reference parameters are useful for describing objective weights for each pair of MG and NMG. These parameters can be used for multiple objectives to express the physical interpretation of a dynamical system. In addition, γ D ( G ) π ˙ D ( G ) = 1 ( α D ( G ) T ¨ D ( G ) + β D ( G ) S ¨ D ( G ) ) , where π ˙ D ( G ) is called the indeterminacy degree of G to D and γ D ( G ) is the reference parameter related to the indeterminacy part. It can be seen that the tuples ( T ¨ D ( G ) , S ¨ D ( G ) , α D ( G ) , β D ( G ) ) with G Q ¨ are crucial for specifying the LDFS D . Due to this fact, we introduce the new notion of the linear Diophantine fuzzy number (LDFN) denoted as A ¨ D = ( t ˙ D , f ˙ D , α D , β D ) satisfying all the constraints listed above for LDFSs. The collection of all LDFSs in Q ¨ is denoted as D ( Q ¨ ) .
Example 1
(Combination of drugs in medicine for better treatment.). Medicines are chemicals or compounds used to cure, halt, or prevent disease, ease symptoms, or help in the diagnosis of illnesses. Advances in medicines have enabled doctors to cure many diseases and save lives. A combination drug or a fixed-dose combination (FDC) is a medicine that includes two or more active ingredients combined in a single dosage form. For example: aspirin/paracetamol and caffeine is a combination drug for the treatment of pain, especially tension headaches and migraines. Let Q ¨ = { G 1 , G 2 , G 3 , G 4 , G 5 } be the collection of some life-saving drugs. In order to gain a high impact of medicine, two or more drugs can be combined in the preparation of a medicine. If the reference or control parameter is considered as:
α = e x c e l l e n t   i m p a c t   a g a i n s t   i n f e c t i o n   p r o d u c e d   d u r i n g   s u r g e r i e s β = n o   h i g h   i m p a c t   a g a i n s t   i n f e c t i o n   p r o d u c e d   d u r i n g   s u r g e r i e s
then its LDFS is given in Table 2.
According to the quality, variety, and severity of the disease, a physician provides medicine to the subject. The information data can be classified using control parameters. These parameters represent how much that portion is necessary for the treatment, and their grade values describe how much that factor is present in that medicine. If we change the parameter as:
α = E x c e l l e n t   i m p a c t   a g a i n s t   e a r   i n f e c t i o n β = N o t   h i g h l y   a f f e c t i v e   f o r   e a r   i n f e c t i o n O R α = F e w e r   s i d e   e f f e c t s β = M o r e   s i d e   e f f e c t s ,
then we can establish various LDFSs that are suitable in other situations. This model helps a pharmacist/doctor/consultant prescribe the most reliable and suitable medicine to the patient for his/her disease. Moreover, reference or control parameters can be used for the purpose of various alternatives in medicine.
Theorem 1
([54]). LDFSs have a larger valuation space than IFSs and PFSs.
Definition 2
([54]). Let A ¨ D = ( t ˙ D , f ˙ D , α D , β D ) be an LDFN and X > 0 . Then:
  • A ¨ D c = ( f ˙ D , t ˙ D , β D , α D ) ;
  • X A ¨ D = 1 ( 1 t ˙ D ) X , f ˙ D X , 1 ( 1 α D ) X , β D X ;
  • A ¨ D X = t ˙ D X , 1 ( 1 f ˙ D ) X , α D X , 1 ( 1 β D ) X .
Definition 3
([54]). Let A ¨ i = ( t ˙ D i , f ˙ D i , α D , β D ) be two LDFNs with i = 1 , 2 . Then:
  • A ¨ D 1 A ¨ D 2 t ˙ D 1 t ˙ D 2 , f ˙ D 2 f ˙ D 1 , α D 1 α D 2 , β D 2 β D 1 ;
  • A ¨ D 1 = A ¨ D 2 t ˙ D 1 = t ˙ D 2 , f ˙ D 1 = f ˙ D 2 , α D 1 = α D 2 , β D 1 = β D 2 ;
  • A ¨ D 1 A ¨ D 2 = t ˙ D 1 + t ˙ D 2 t ˙ D 1 t ˙ D 2 , f ˙ D 1 f ˙ D 2 , α D 1 + α D 2 α D 1 α D 2 , β D 1 β D 2 ;
  • A ¨ D 1 A ¨ 2 = t ˙ D 1 t ˙ D 2 , f ˙ D 1 + f ˙ D 2 f ˙ D 1 f ˙ D 2 , α D 1 α D 2 , β D 1 + β D 2 β D 1 β D 2 .
Definition 4
([54]). Let A ¨ D i = ( t ˙ D i , f ˙ D i , α D i , β D i ) be a collection of LDFNs with i Δ . Then:
  • i Δ A ¨ D i = sup i Δ t ˙ D i , inf i Δ f ˙ D i , sup i Δ α D i , inf i Δ β D i ;
  • i Δ A ¨ D i = inf i Δ t ˙ D i , sup i Δ f ˙ D i , inf i Δ α D i , sup i Δ β D i .
Definition 5
([56]). For the non-empty collection of alternatives Q ¨ and the collection of attributes G ˙ , the soft set is evaluated by the mapping Ω ˙ : G ˙ P ˜ ( Q ¨ ) . Alternatively, it can be represented as:
( Ω ˙ , G ˙ ) = { ( , Ω ˙ ( ) ) : Ω ˙ ( ) P ˜ ( Q ¨ ) , G ˙ }
The collection of all subsets of Q ¨ is denoted as P ˜ ( Q ¨ ) .
Definition 6
([55]). Suppose the indiscernibility relation on Q ¨ is denoted as R . We assume arbitrarily that R is an equivalence relation. Moreover, N e g R K = Q ¨ K , P o s R K = K , and B n d R K = K K are said to be negative, positive, and boundary regions of K Q ¨ . The characteristics of these regions are given as follows:
(1)
G P o s R K implies that K certainly contains the elements G of Q ¨ .
(2)
G N e g R K implies that K does not contains the elements G of Q ¨ .
(3)
G B n d R K implies that K may or may not contain the elements G of Q ¨ .
The equivalence class of object G under the relation R is represented as [ G ] R . The pair ( Q ¨ , R ) is said to be a “Pawlak approximation space”, and R will generate the partition Q ¨ / R = { [ G ] R : G Q ¨ } . Then, pair ( R ( K ) , R ( K ) ) is called the rough set of crisp set K , where:
R ( K ) = { G Q ¨ : [ G ] R K }
R ( K ) = { G Q ¨ : [ G ] R K ϕ }
are called “lower and upper approximations” of K with respect to ( Q ¨ , R ) . If R ( K ) = R ( K ) , then K is said to be definable; otherwise, it is called a rough set.
Remark 1.
The concepts of the core and reduct in rough set theory are very significant tools in the decision making methods. We can deduce the reduct from the reference set Q ¨ . It is used to reduce the unimportant information in the input data. The core is the intersection of all reducts and provides the final optimal decision about the decision making problem (see [3,32]).
Definition 7
([26]). For a non-empty collection of alternatives Q ¨ and the collection of attributes G ˙ , the crisp soft relation R Q ¨ × G ˙ is written as:
R = { ( G , ˙ ) , ψ R ( G , ˙ ) : ( G , ˙ ) Q ¨ × G ˙ }
where ψ R : Q ¨ × G ˙ { 0 , 1 } and:
ψ R ( G , ˙ ) = 1 i f ( G , ˙ ) R 0 ; o t h e r w i s e
Definition 8
([26]). For a non-empty collection of alternatives Q ¨ and the collection of attributes G ˙ , we have a crisp soft relation A ˜ Q ¨ × G ˙ . A mapping A ˜ s : Q ¨ P ( G ˙ ) is written as:
A ˜ s ( G ) = { ˙ G ˙ : ( G , ˙ ) A ˜ ; G Q ¨ }
A ˜ is called serial if G Q ¨ , A ˜ s ( G ) ϕ . The “crisp soft approximation space” is represented by this triplet ( Q ¨ , G ˙ , A ˜ ) . For arbitrary H G ˙ , A ˜ ( H ) and A ˜ ( H ) are called the “lower and upper approximations”, respectively, defined as:
A ˜ ( H ) = { G Q ¨ : A ˜ s ( G ) H ϕ }
A ˜ ( H ) = { G Q ¨ : A ˜ s ( G ) Q ¨ }
The pair ( A ˜ ( H ) , A ˜ ( H ) ) is called the crisp soft rough set, and A ˜ , A ˜ : P ˜ ( G ˙ ) P ˜ ( Q ¨ ) are called “lower and upper approximation operators”. P ˜ ( G ˙ ) and P ˜ ( Q ¨ ) are an assembly of all subsets of G ˙ and Q ¨ , respectively. If A ˜ ( H ) = A ˜ ( H ) , then Q ¨ is called definable.

3. Construction of SRLDFSs and LDFSRSs

In this part, we organize the innovative hybrid structures of soft rough linear Diophantine fuzzy sets (SRLDFSs) and linear Diophantine fuzzy soft rough sets (LDFSRSs) by merging the fundamental compositions of LDFSs, soft sets, and rough sets. In decision making obstacles, we deal with the ambiguities and vagueness in the initial input information. Due to these circumstances, we cannot manage these inputs by utilizing simplistic models. In fuzzy sets, IFSs, PFSs, and q-ROFSs, the opportunities for the assortment of satisfaction and dissatisfaction degrees are restricted due to constraints 0 T ˙ 1 , 0 T ¨ + S ¨ 1 , 0 T ¨ 2 + S ¨ 2 1 , and 0 T ¨ q + S ¨ q 1 . However, in the LDFS, we can comfortably choose the degrees from [ 0 , 1 ] , due to the reference or control parameters. However, this set does not deal with the vagueness or roughness. We cannot handle uncertainties and parameterizations if we deal only with the roughness of a set. The soft set only works for parameterizations. Therefore, to eliminate these ambiguities and to fill in the research gap, we assemble SRLDFSs and LDFSRSs. These models dispense with the fuzzy degrees, parameterizations, and roughness of the data in the decision making difficulties. The significance of these generalized and authentic notions can be examined in the entire article. Table 3 represents the notations used in the whole manuscript.

3.1. Soft Rough Linear Diophantine Fuzzy Sets

Definition 9.
For the reference set Q ¨ and set of decision variables G ˙ , if we define a crisp soft relation A ˜ over Q ¨ × G ˙ , then ( Q ¨ , G ˙ , A ˜ ) is called a “crisp soft approximation space”. If Y D D ( G ˙ ) , then A ˜ ( Y D ) and A ˜ ( Y D ) are called “upper and lower approximations” of Y D about ( Q ¨ , G ˙ , A ˜ ) respectively and written as:
A ˜ ( Y D ) = { ( G , T ¨ A ˜ ( Y D ) ( G ) , S ¨ A ˜ ( Y D ) ( G ) , α A ˜ ( Y D ) ( G ) , β A ˜ ( Y D ) ( G ) ) : G Q ¨ }
A ˜ ( Y D ) = { ( G , T ¨ A ˜ ( Y D ) ( G ) , S ¨ A ˜ ( Y D ) ( G ) , α A ˜ ( Y D ) ( G ) , β A ˜ ( Y D ) ( G ) ) : G Q ¨ }
where the degrees can be calculated as given in Table 4.
The notions given in Table 4 satisfy the following constraints:
0 α A ˜ ( Y D ) ( G ) T ¨ A ˜ ( Y D ) ( G ) + β A ˜ ( Y D ) ( G ) S ¨ A ˜ ( Y D ) ( G ) 1
0 α A ˜ ( Y D ) ( G ) T ¨ A ˜ ( Y D ) ( G ) + β A ˜ ( Y D ) ( G ) S ¨ A ˜ ( Y D ) ( G ) 1
0 α A ˜ ( Y D ) ( G ) + β A ˜ ( Y D ) ( G ) 1 a n d
0 α A ˜ ( Y D ) ( G ) + β A ˜ ( Y D ) ( G ) 1
D ( G ˙ ) is an assembly of LDFSs over G ˙ . A ˜ ( Y D ) and A ˜ ( Y D ) are LDFSs over Q ¨ . Thus, the pair ( A ˜ ( Y D ) , A ˜ ( Y D ) ) is called the soft rough linear Diophantine fuzzy set (SRLDFS) about ( Q ¨ , G ˙ , A ˜ ) , and A ˜ , A ˜ : D ( G ˙ ) D ( Q ¨ ) are called upper and lower SRLDF approximation operators. If A ˜ ( Y D ) = A ˜ ( Y D ) , then Y D is called definable.
Example 2.
We consider the collection of well known cars given as Q ¨ = { G 1 , G 2 , G 3 , G 4 } and the assembly of suitable attributes G ˙ = { ˙ 1 , ˙ 2 , ˙ 3 , ˙ 4 } . The attributes are given as “comfortable and reliable”, “good safety”, “good maintenance”, and “affordable”. Let ( η , G ˙ ) be the soft set in Q ¨ given as:
η ( ˙ 1 ) = { G 1 , G 2 , G 3 } , η ( ˙ 2 ) = { G 2 , G 4 } η ( ˙ 3 ) = { G 1 , G 2 , G 3 , G 4 } , η ( ˙ 4 ) = { G 1 , G 4 }
A crisp relation over Q ¨ × G ˙ is given as A ˜ = { ( G 1 , ˙ 1 ) , ( G 2 , ˙ 1 ) , ( G 3 , ˙ 1 ) , ( G 2 , ˙ 2 ) , ( G 4 , ˙ 2 ) , ( G 1 , ˙ 3 ) , ( G 2 , ˙ 3 ) , ( G 3 , ˙ 3 ) , ( G 4 , ˙ 3 ) , ( G 1 , ˙ 4 ) , ( G 4 , ˙ 4 ) } . By definition, we have:
A ˜ s ( G 1 ) = { ˙ 1 , ˙ 3 , ˙ 4 } A ˜ s ( G 2 ) = { ˙ 1 , ˙ 2 , ˙ 3 } A ˜ s ( G 3 ) = { ˙ 1 , ˙ 3 } A ˜ s ( G 4 ) = { ˙ 2 , ˙ 3 , ˙ 4 }
We consider LDFS, Y D D ( G ˙ ) , given as:
Y D = { ( ˙ 1 , 0.786 , 0.765 , 0.234 , 0.123 ) , ( ˙ 2 , 0.987 , 0.574 , 0.232 , 0.423 ) , ( ˙ 3 , 0.912 , 0.536 , 0.235 , 0.635 ) , ( ˙ 4 , 0.726 , 0.825 , 0.765 , 0.122 ) }
The “upper and lower approximations” can be computed by using Definition 9. Upper approximations are given as:
T ¨ A ˜ ( Y D ) ( G 1 ) = 0.912 , S ¨ A ˜ ( Y D ) ( G 1 ) = 0.536 , α A ˜ ( Y D ) ( G 1 ) = 0.765 , β A ˜ ( Y D ) ( G 1 ) = 0.122 T ¨ A ˜ ( Y D ) ( G 2 ) = 0.987 , S ¨ A ˜ ( Y D ) ( G 2 ) = 0.574 , α A ˜ ( Y D ) ( G 2 ) = 0.765 , β A ˜ ( Y D ) ( G 2 ) = 0.122 T ¨ A ˜ ( Y D ) ( G 3 ) = 0.912 , S ¨ A ˜ ( Y D ) ( G 3 ) = 0.536 , α A ˜ ( Y D ) ( G 3 ) = 0.235 , β A ˜ ( Y D ) ( G 3 ) = 0.123 T ¨ A ˜ ( Y D ) ( G 4 ) = 0.987 , S ¨ A ˜ ( Y D ) ( G 4 ) = 0.536 , α A ˜ ( Y D ) ( G 4 ) = 0.765 , β A ˜ ( Y D ) ( G 4 ) = 0.122
Lower approximations are evaluated as:
T ¨ A ˜ ( Y D ) ( G 1 ) = 0.726 , S ¨ A ˜ ( Y D ) ( G 1 ) = 0.825 , α A ˜ ( Y D ) ( G 1 ) = 0.234 , β A ˜ ( Y D ) ( G 1 ) = 0.635 T ¨ A ˜ ( Y D ) ( G 2 ) = 0.786 , S ¨ A ˜ ( Y D ) ( G 2 ) = 0.765 , α A ˜ ( Y D ) ( G 2 ) = 0.232 , β A ˜ ( Y D ) ( G 2 ) = 0.635 T ¨ A ˜ ( Y D ) ( G 3 ) = 0.786 , S ¨ A ˜ ( Y D ) ( G 3 ) = 0.765 , α A ˜ ( Y D ) ( G 3 ) = 0.234 , β A ˜ ( Y D ) ( G 3 ) = 0.635 T ¨ A ˜ ( Y D ) ( G 4 ) = 0.726 , S ¨ A ˜ ( Y D ) ( G 4 ) = 0.825 , α A ˜ ( Y D ) ( G 4 ) = 0.232 , β A ˜ ( Y D ) ( G 4 ) = 0.635
Thus:
A ˜ ( Y D ) = { ( G 1 , 00.912 , 00.536 , 00.765 , 00.122 ) , ( G 2 , 00.987 , 00.574 , 00.765 , 00.122 ) , ( G 3 , 00.912 , 00.536 , 00.235 , 00.123 ) } , ( G 4 , 00.987 , 00.536 , 00.765 , 00.122 ) } A ˜ ( Y D ) = { ( G 1 , 00.726 , 00.825 , 00.234 , 00.635 ) , ( G 2 , 00.786 , 00.765 , 00.232 , 00.635 ) , ( G 3 , 00.786 , 00.765 , 00.234 , 00.635 ) } , ( G 4 , 00.726 , 00.825 , 00.232 , 00.635 ) }
Therefore, ( A ˜ ( Y D ) , A ˜ ( Y D ) ) is said to be SRLDFS.
Remark 2.
For the “crisp soft approximation space” ( Q ¨ , G ˙ , A ˜ ) , if we take the upper and lower approximations of the following sets listed in Table 5, then we can observe the degeneration of SRLDF approximation operators into different structures based on rough sets.
It is evident from Table 5 that our proposed model is superior and powerful in contrast with other existing structures. However, we cannot decompose the described theories into the SRLDFSs and their respective approximation operators. In simple terms, SRLDFS is the generalization of “soft rough sets, soft rough fuzzy sets, soft rough intuitionistic fuzzy sets, soft rough Pythagorean fuzzy sets, and soft rough q-rung orthopair fuzzy sets”.
Theorem 2.
Let Y D , B D D ( G ) and A ˜ ( Y D ) , A ˜ ( Y D ) be “upper and lower approximation operators” over the approximation space ( Q ¨ , G ˙ , A ˜ ) , then the following axioms are true:
(1) 
A ˜ ( Y D ) = A ˜ ( Y D ) ,
(2) 
Y D B D A ˜ ( Y D ) A ˜ ( B D ) ,
(3) 
A ˜ ( Y D B D ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(4) 
A ˜ ( Y D B D ) A ˜ ( Y D ) A ˜ ( B D ) ,
(5) 
A ˜ ( Y D ) = A ˜ ( Y D ) ,
(6) 
Y D B D A ˜ ( Y D ) A ˜ ( B D ) ,
(7) 
A ˜ ( Y D B D ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(8) 
A ˜ ( Y D B D ) A ˜ ( Y D ) A ˜ ( B D ) .
The complement of Y D is represented by Y D .
Proof. 
See Appendix A. □
Now, we provide a counter example to prove that equality does not exist in Parts (4) and (8) of Theorem 2.
Example 3.
For the reference set Q ¨ = { G 1 , G 2 , G 3 , G 4 } and assembly of decision variables G ˙ = { ˙ 1 , ˙ 2 , ˙ 3 } , we define a soft set ( η , G ˙ ) in Q ¨ written as:
β ( ˙ 1 ) = { G 1 , G 4 } , β ( ˙ 2 ) = { G 1 , G 2 , G 4 } , β ( ˙ 3 ) = { G 2 , G 3 }
The crisp soft relation A ˜ in Q ¨ × G ˙ is given as A ˜ = { ( G 1 , ˙ 1 ) , ( G 4 , ˙ 1 ) , ( G 1 , ˙ 2 ) , ( G 2 , ˙ 2 ) , ( G 4 , ˙ 2 ) , ( G 2 , ˙ 3 ) , ( G 3 , ˙ 3 ) } . We can write it as:
A ˜ s ( G 1 ) = { ˙ 1 , ˙ 2 } , A ˜ s ( G 2 ) = { ˙ 2 , ˙ 3 } , A ˜ s ( G 3 ) = { ˙ 3 } , A ˜ s ( G 4 ) = { ˙ 1 , ˙ 2 }
Let Y D , B D D ( G ) be given as follows:
Y D = { ( ˙ 1 , 0.573 , 0.273 , 0.271 , 0.531 ) , B D = { ( ˙ 1 , 0.773 , 0.273 , 0.281 , 0.523 ) , ( ˙ 2 , 0.378 , 0.177 , 0.291 , 0.532 ) , ( ˙ 2 , 0.778 , 0.371 , 0.283 , 0.521 ) , ( ˙ 3 , 0.678 , 0.178 , 0.271 , 0.521 ) } ( ˙ 3 , 0.873 , 0.371 , 0.261 , 0.532 ) }
The “upper approximations” are given as:
A ˜ ( Y D ) = { ( G 1 , 00.573 , 00.177 , 00.291 , 00.531 ) , ( G 2 , 00.678 , 00.177 , 00.291 , 00.521 ) , ( G 3 , 00.678 , 00.178 , 00.271 , 00.521 ) , ( G 4 , 00.573 , 00.177 , 00.291 , 00.531 ) } A ˜ ( B D ) = { ( G 1 , 00.778 , 00.273 , 00.283 , 00.521 ) , ( G 2 , 00.873 , 00.371 , 00.261 , 00.521 ) , ( G 3 , 00.873 , 00.371 , 00.261 , 00.532 ) , ( G 4 , 00.778 , 00.273 , 00.283 , 00.521 ) } A ˜ ( Y D B D ) = { ( G 1 , 00.573 , 00.273 , 00.283 , 00.531 ) , ( G 2 , 00.678 , 00.371 , 00.261 , 00.532 ) , ( G 3 , 00.678 , 00.371 , 00.261 , 00.532 ) , ( G 4 , 00.573 , 00.273 , 00.283 , 00.531 ) } A ˜ ( Y D ) A ˜ ( B D ) = { ( G 1 , 00.573 , 00.273 , 00.283 , 00.531 ) , ( G 2 , 00.678 , 00.371 , 00.261 , 00.521 ) , ( G 3 , 00.678 , 00.371 , 00.261 , 00.532 ) , ( G 4 , 00.573 , 00.273 , 00.283 , 00.531 ) }
From the above calculations, it is clear that A ˜ ( Y D ) A ˜ ( B D ) A ˜ ( Y D B D ) as for alternative G 2 , the degrees of reference parameter β A ˜ ( Y D ) A ˜ ( B D ) ( G 2 ) β A ˜ ( Y D B D ) ( G 2 ) , i.e., 00.521 00.532 . Similarly, we can check that A ˜ ( Y D B D ) A ˜ ( Y D ) A ˜ ( B D ) .
Proposition 1.
If Y D , B D D ( G ) , then A ˜ ( Y D ) , A ˜ ( B D ) , A ˜ ( Y D ) and A ˜ ( B D ) are “lower and upper approximations” of LDFSs over the “crisp soft approximation space” ( Q ¨ , G ˙ , A ˜ ) satisfying the following axioms:
(1) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(2) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(3) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(4) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(5) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(6) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(7) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) ,
(8) 
( A ˜ ( Y D ) A ˜ ( B D ) ) = A ˜ ( Y D ) A ˜ ( B D ) .
Proof. 
The proof is obvious. □

3.2. Linear Diophantine Fuzzy Soft Rough Sets

Definition 10.
For the non-empty set of alternatives Q ¨ and the collection of attributes G ˙ , we consider a subset O ˙ G ˙ . Then, we define linear Diophantine fuzzy soft set (LDFSS), ( δ ¨ , O ˙ ) represented by the mapping:
δ ¨ : O ˙ D ( Q ¨ )
where D ( Q ¨ ) is an assembly of all LDF-subsets of Q ¨ . Alternatively, it can be written as:
( δ ¨ , O ˙ ) = { ( ˙ , δ ¨ ( ˙ ) ) : ˙ O ˙ , δ ¨ ( ˙ ) D ( Q ¨ ) }
Definition 11.
Let ( δ ¨ , O ˙ ) be an LDFSS in Q ¨ . Then, an LDF-subset ð ˜ of Q ¨ × G ˙ is called a linear Diophantine fuzzy soft relation (LDFSR) from Q ¨ to G ˙ written as:
ð ˜ = { ( ( G , ˙ ) , T ¨ ð ˜ ( G , ˙ ) , S ¨ ð ˜ ( G , ˙ ) , α ð ˜ ( G , ˙ ) , β ð ˜ ( G , ˙ ) ) : ( G , ˙ ) Q ¨ × G ˙ }
where α T ¨ ð ˜ ( G , ˙ ) , α S ¨ ð ˜ ( G , ˙ ) [ 0 , 1 ] are truth and falsity grades, respectively, with the corresponding reference parameters α ð ˜ ( G , ˙ ) , β ð ˜ ( G , ˙ ) [ 0 , 1 ] satisfying the constraints:
0 α ð ˜ ( G , ˙ ) α T ¨ ð ˜ 2 ( G , ˙ ) + β ð ˜ ( G , ˙ ) α S ¨ ð ˜ 2 ( G , ˙ ) 1
0 α ð ˜ ( G , ˙ ) + β ð ˜ ( G , ˙ ) 1
If Q ¨ = { G 1 , G 2 , . . . , G n } and G ˙ = { ˙ 1 , ˙ 2 , . . . , ˙ m } , then LDFSR ð ˜ on Q ¨ × G ˙ can be represented in tabular form as Table 6.
Definition 12.
For the reference set Q ¨ and set of decision variables G ˙ , if we define an LDFSR ð ˜ over Q ¨ × G ˙ , then ( Q ¨ , G ˙ , ð ˜ ) is called an “LDFS approximation space”. If Y D D ( G ˙ ) , then ð ˜ ( Y D ) and ð ˜ ( Y D ) are “upper and lower approximations” of Y D about ( Q ¨ , G ˙ , ð ˜ ) respectively and written as:
ð ˜ ( Y D ) = { ( G , T ¨ ð ˜ ( Y D ) ( G ) , S ¨ ð ˜ ( Y D ) ( G ) , α ð ˜ ( Y D ) ( G ) , β ð ˜ ( Y D ) ( G ) ) : G Q ¨ }
ð ˜ ( Y D ) = { ( G , T ¨ ð ˜ ( Y D ) ( G ) , S ¨ ð ˜ ( Y D ) ( G ) , α ð ˜ ( Y D ) ( G ) , β ð ˜ ( Y D ) ( G ) ) : G Q ¨ }
where the degrees can be calculated as given in Table 7.
The pair ( ð ˜ ( Y D ) , ð ˜ ( Y D ) ) is a called linear Diophantine fuzzy soft rough set (LDFSRS) in ( Q ¨ , G ˙ , ð ˜ ) . The “lower and upper approximation operators” are represented as ð ˜ ( Y D ) and ð ˜ ( Y D ) , respectively. If ð ˜ ( Y D ) = ð ˜ ( Y D ) , then Y D is said to be definable.
Example 4.
Let Q ¨ = { G 1 , G 2 } be the collection of certain cloth brands and G ˙ = { ˙ 1 , ˙ 2 , ˙ 3 } be the set of attributes, where:
˙ 1 = P r o d u c t   q u a l i t y , ˙ 2 = A f f o r d a b l e , ˙ 3 = R e c o v e r y   s e r v i c e .
We construct the LDFSR, ð ˜ : Q ¨ G ˙ , represented in Table 8.
Consider a linear Diophantine fuzzy soft subset Y D of G ˙ given as:
Y D = { ( ˙ 1 , 0.837 , 0.535 , 0.242 , 0.242 ) , ( ˙ 2 , 0.833 , 0.635 , 0.634 , 0.142 ) , ( ˙ 3 , 0.725 , 0.526 , 0.625 , 0.211 ) }
By using Definition 12, we find the “upper and lower approximations” of Y D given by:
T ¨ ð ˜ ( Y D ) ( G 1 ) = ˙ [ 0.684 , 0.825 , 0.725 ] = 0.825 , S ¨ ð ˜ ( Y D ) ( G 1 ) = max ˙ [ 0.645 , 0.635 , 0.735 ] = 0.635 , α ð ˜ ( Y D ) ( G 1 ) = max ˙ [ 0.221 , 0.226 , 0.122 ] = 0.226 , β ð ˜ ( Y D ) ( G 1 ) = min ˙ [ 0.675 , 0.877 , 0.677 ] = 0.675
Similarly, we find all other values for the “upper and lower approximation” of Y D . This implies that:
ð ˜ ( Y D ) = { ( G 1 , 0.825 , 0.635 , 0.226 , 0.675 ) , ( G 2 , 0.837 , 0.535 , 0.242 , 0.348 ) } ð ˜ ( Y D ) = { ( G 1 , 0.725 , 0.635 , 0.774 , 0.242 ) , ( G 2 , 0.752 , 0.635 , 0.754 , 0.242 ) }
Thus, ( ð ˜ ( Y D ) , ð ˜ ( Y D ) ) is called LDFSRS.
Remark 3.
For the “linear Diophantine fuzzy soft approximation space (LDFS approximation space)” ( Q ¨ , G ˙ , ð ˜ ) , if we take the upper and lower approximations of the following sets listed in Table 9, then we can observe the degeneration of LDFSR approximation operators into different structures based on rough sets.
It is evident from Table 9 that our proposed model is superior and powerful in contrast with other existing structures. However, we cannot decompose the described theories into the LDFSRSs and their respective approximation operators. The beauty of this structure is that if we select the “crisp soft approximation space” for LDFSR approximation operators, then it will be degenerated into the proposed SRLDFSs. This generalization provides us a strong relation between both proposed rough set models. In simple terms, LDFSRS is the generalization of “soft fuzzy rough sets, intuitionistic fuzzy soft rough sets, Pythagorean fuzzy soft rough sets, q-rung orthopair fuzzy soft rough sets, and soft rough linear Diophantine fuzzy sets”.
Theorem 3.
For arbitrary Y D , B D D ( G ) , the “upper and lower approximation operators” ð ˜ ( Y D ) , ð ˜ ( B D ) , ð ˜ ( Y D ) and ð ˜ ( B D ) on the “LDFS approximation space” ( Q ¨ , G ˙ , ð ˜ ) satisfy the following axioms:
(1) 
ð ˜ ( Y D ) = ð ˜ ( Y D ) ,
(2) 
Y D B D ð ˜ ( Y D ) ð ˜ ( B D ) ,
(3) 
ð ˜ ( Y D B D ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(4) 
ð ˜ ( Y D B D ) ð ˜ ( Y D ) ð ˜ ( B D ) ,
(5) 
ð ˜ ( Y D ) = ð ˜ ( Y D ) ,
(6) 
Y D B D ð ˜ ( Y D ) ð ˜ ( B D ) ,
(7) 
ð ˜ ( Y D B D ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(8) 
ð ˜ ( Y D B D ) ð ˜ ( Y D ) ð ˜ ( B D ) .
The complement of Y D is represented by Y D .
Proof. 
The proof is similar to the proof given in Appendix A. □
Proposition 2.
For arbitrary Y D , B D D ( G ) , the “upper and lower approximation operators” ð ˜ ( Y D ) , ð ˜ ( B D ) , ð ˜ ( Y D ) and ð ˜ ( B D ) on the “LDFS approximation space” ( Q ¨ , G ˙ , ð ˜ ) satisfy the following axioms:
(1) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(2) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(3) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(4) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(5) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(6) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(7) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) ,
(8) 
( ð ˜ ( Y D ) ð ˜ ( B D ) ) = ð ˜ ( Y D ) ð ˜ ( B D ) .
Proof. 
The proof is obvious. □
Theorem 4.
For “LDFS approximation space” ( Q ¨ , G ˙ , ð ˜ ) , if ð ˜ is serial, then ð ˜ ( Y D ) and ð ˜ ( Y D ) satisfy the following:
(1) 
ð ˜ ( ) = , ð ˜ ( G ˙ ) = G ˙ ,
(2) 
ð ˜ ( Y D ) ð ˜ ( Y D ) , Y D Y D D ( G ) .
Proof. 
The proof is obvious by following Definition 12. □
Definition 13.
Let Y D D ( Q ¨ ) , and let ð ˜ ( Y D ) , ð ˜ ( Y D ) be lower and upper “LDFSR approximation operators”. Then, the ring sum operation of ð ˜ ( Y D ) and ð ˜ ( Y D ) is written as:
ð ˜ ( Y D ) ð ˜ ( Y D ) = { ( G , T ¨ ð ˜ ( Y D ) ( G ) + T ¨ ð ˜ ( Y D ) ( G ) ( T ¨ ð ˜ ( Y D ) ( G ) × T ¨ ð ˜ ( Y D ) ( G ) ) , S ¨ ð ˜ ( Y D ) ( G ) × S ¨ ð ˜ ( Y D ) ( G ) , α ð ˜ ( Y D ) ( G ) + α ð ˜ ( Y D ) ( G ) ( α ð ˜ ( Y D ) ( G ) × α ð ˜ ( Y D ) ( G ) ) , β ð ˜ ( Y D ) ( G ) × β ð ˜ ( Y D ) ( G ) ) : G Q ¨ }
Definition 14.
Let D = G , T ¨ D ( G ) , S ¨ D ( G ) , α D ( G ) , β D ( G ) : G Q ¨ be an LDFS and the constants ( η , θ , ζ , ψ ) , where η , θ , ζ , ψ [ 0 , 1 ] satisfying the constraints 0 η θ + ζ ψ 1 and 0 θ + ψ 1 . Then, the ( η , θ , ζ , ψ ) -level cut set of D is written as:
D η , θ ζ , ψ = { G Q ¨ : T ¨ D ( G ) η , α D ( G ) θ , S ¨ D ( G ) ζ , β D ( G ) ψ } .
The η , θ -level cut of D is written as D η , θ = { G Q ¨ : T ¨ D ( G ) η , α D ( G ) θ } .
The strong η , θ -level cut of D is written as D η , θ + = { G Q ¨ : T ¨ D ( G ) > η , α D ( G ) > θ } .
The ζ , ψ -level cut of D is written as D ζ , ψ = { G Q ¨ : S ¨ D ( G ) ζ , β D ( G ) ψ } .
The strong ζ , ψ -level cut of D is written as D ζ , ψ + = { G Q ¨ : S ¨ D ( G ) < ζ , β D ( G ) < ψ } .
The other cut sets of an LDFS are analogously described as:
D η , θ + ζ , ψ = { G Q ¨ : T ¨ D ( G ) > η , α D ( G ) > θ , S ¨ D ( G ) ζ , β D ( G ) ψ } .
D η , θ ζ , ψ + = { G Q ¨ : T ¨ D ( G ) η , α D ( G ) θ , S ¨ D ( G ) < ζ , β D ( G ) < ψ } .
D η , θ + ζ , ψ + = { G Q ¨ : T ¨ D ( G ) > η , α D ( G ) > θ , S ¨ D ( G ) < ζ , β D ( G ) < ψ } .
Theorem 5.
Let D , 1 D , 2 D D ( Q ¨ ) and η , θ , ζ , ψ [ 0 , 1 ] satisfy the constraints 0 η θ + ζ ψ 1 and 0 θ + ψ 1 . Then, the cut sets of LDFSs satisfy the following axioms:
1
D η , θ ζ , ψ = D η , θ D ζ , ψ ,
2
( D ) η , θ = D η , θ + , ( D ) ζ , ψ = D ζ , ψ + ,
3
1 D 2 D = 1 D η , θ ζ , ψ 2 D η , θ ζ , ψ ,
4
( 1 D 2 D ) η , θ = 1 D η , θ 2 D η , θ , ( 1 D 2 D ) ζ , ψ = 1 D ζ , ψ 2 D ζ , ψ ,
( 1 D 2 D ) η , θ ζ , ψ = 1 D η , θ ζ , ψ 2 D η , θ ζ , ψ ,
5
( 1 D 2 D ) η , θ = 1 D η , θ 2 D η , θ , ( 1 D 2 D ) ζ , ψ = 1 D ζ , ψ 2 D ζ , ψ ,
( 1 D 2 D ) η , θ ζ , ψ = 1 D η , θ ζ , ψ 2 D η , θ ζ , ψ
6
If η 1 η 2 , θ 1 θ 2 and ζ 1 ζ 2 , ψ 1 ψ 2 , then
D η 1 , θ 1 D η 2 , θ 2 , D ζ 1 , ψ 1 D ζ 2 , ψ 2 and D η 1 , θ 1 ζ 1 , ψ 1 D η 1 , θ 1 ζ 2 , ψ 2 .
Proof. 
This proof is inferred explicitly by Definition 14. □
By using the defined idea of cut sets on LDFSs, we can find the cut sets of LDFSR:
ð ˜ = { ( ( G , ˙ ) , T ¨ ð ˜ ( G , ˙ ) , S ¨ ð ˜ ( G , ˙ ) , α ð ˜ ( G , ˙ ) , β ð ˜ ( G , ˙ ) ) : ( G , ˙ ) Q ¨ × G ˙ }
given as:
ð ˜ η , θ = { ( ( G , ˙ ) Q ¨ × G ˙ : T ¨ ð ˜ ( G , ˙ ) η , α ð ˜ ( G , ˙ ) θ }
ð ˜ η , θ ( G ) = { ˙ G ˙ : T ¨ ð ˜ ( G , ˙ ) η , α ð ˜ ( G , ˙ ) θ } f o r η , θ [ 0 , 1 ]
ð ˜ η , θ + = { ( ( G , ˙ ) Q ¨ × G ˙ : T ¨ ð ˜ ( G , ˙ ) > η , α ð ˜ ( G , ˙ ) > θ }
ð ˜ η , θ + ( G ) = { ˙ G ˙ : T ¨ ð ˜ ( G , ˙ ) > η , α ð ˜ ( G , ˙ ) > θ } f o r η , θ [ 0 , 1 )
ð ˜ η , θ = { ( ( G , ˙ ) Q ¨ × G ˙ : S ¨ ð ˜ ( G , ˙ ) η , β ð ˜ ( G , ˙ ) θ }
ð ˜ η , θ ( G ) = { ˙ G ˙ : S ¨ ð ˜ ( G , ˙ ) η , β ð ˜ ( G , ˙ ) θ } f o r η , θ [ 0 , 1 ]
ð ˜ η , θ + = { ( ( G , ˙ ) Q ¨ × G ˙ : S ¨ ð ˜ ( G , ˙ ) < η , β ð ˜ ( G , ˙ ) < θ }
ð ˜ η , θ + ( G ) = { ˙ G ˙ : S ¨ ð ˜ ( G , ˙ ) < η , β ð ˜ ( G , ˙ ) < θ } f o r η , θ ( 0 , 1 ]
where all the calculated cuts are crisp soft relations. Now, we present a result to show that LDFSR approximation operators can be written as crisp soft rough approximation operators.
Theorem 6.
Consider that for LDFSR approximation space ( Q ¨ , G ˙ , ð ˜ ) and D D ( Q ¨ ) , the upper approximation operators can be represented as:
1
T ¨ ð ˜ ( D ) ( G ) , α ð ˜ ( D ) ( G ) = η , θ [ 0 , 1 ] [ η , θ ð ˜ η , θ ( D η , θ ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ η , θ ( D η , θ + ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ η , θ + ( D η , θ ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ η , θ + ( D η , θ + ) ( G ) ]
2
S ¨ ð ˜ ( D ) ( G ) , β ð ˜ ( D ) ( G ) = η , θ [ 0 , 1 ] [ η , θ ( 1 ð ˜ 1 η , 1 θ ( D η , θ ) ( G ) ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ ( D η , θ + ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ + ( D η , θ ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ + ( D η , θ + ) ( G ) ]
and for arbitrary η , θ [ 0 , 1 ] , we have:
3
[ ð ˜ ( D ) ] η , θ + ð ˜ η , θ + ( D η , θ + ) ð ˜ η , θ + ( D η , θ ) ð ˜ η , θ ( D η , θ ) [ ð ˜ ( D ) ] η , θ .
4
[ ð ˜ ( D ) ] η , θ + ð ˜ 1 η , 1 θ + ( D η , θ + ) ð ˜ 1 η , 1 θ + ( D η , θ ) ð ˜ 1 η , 1 θ ( D η , θ ) [ ð ˜ ( D ) ] η , θ .
Proof. 
One can conclude the proof of this theorem directly by using Definitions 12 and 14. □
Theorem 7.
Consider that for LDFSR approximation space ( Q ¨ , G ˙ , ð ˜ ) and D D ( Q ¨ ) , the upper approximation operators can be represented as:
1
T ¨ ð ˜ ( D ) ( G ) , α ð ˜ ( D ) ( G ) = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ ( D η , θ + ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ + ( D η , θ ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ + ( D η , θ + ) ( G ) ] = η , θ [ 0 , 1 ] [ η , θ ð ˜ 1 η , 1 θ ( D η , θ ) ( G ) ]
2
S ¨ ð ˜ ( D ) ( G ) , β ð ˜ ( D ) ( G ) = η , θ [ 0 , 1 ] [ η , θ ( 1 ð ˜ η , θ ( D η , θ ) ( G ) ) ] = η , θ [ 0 , 1 ] [ η , θ ( 1 ð ˜ η , θ + ( D η , θ ) ( G ) ) ] = η , θ [ 0 , 1 ] [ η , θ ( 1 ð ˜ η , θ + ( D η , θ + ) ( G ) ) ] = η , θ [ 0 , 1 ] [ η , θ ( 1 ð ˜ η , θ ( D η , θ + ) ( G ) ) ]
and for arbitrary η , θ [ 0 , 1 ] , we have:
3
[ ð ˜ ( D ) ] η , θ + ð ˜ 1 η , 1 θ ( D η , θ + ) ð ˜ 1 η , 1 θ + ( D η , θ + ) ð ˜ 1 η , 1 θ ( D η , θ ) [ ð ˜ ( D ) ] η , θ .
4
[ ð ˜ ( D ) ] η , θ + ð ˜ 1 η , 1 θ + ( D η , θ + ) ð ˜ 1 η , 1 θ + ( D η , θ ) ð ˜ 1 η , 1 θ ( D η , θ ) [ ð ˜ ( D ) ] η , θ .
Proof. 
The proof of this theorem can be obtained directly by using Definitions 12 and 14. □

4. MCDM for Sustainable Material Handling Equipment

The determination of material handling equipment is extremely substantial in the project of an operative industrial system. The efficiency of material flow depends on the selection of appropriate material handling equipment. It promotes capability utilization and increases productivity. Decision support systems and various programs have been developed by various researchers for the selection of the best material handling equipment. In this section, we establish the novel methodologies for the selection of the appropriate and most reliable material handling equipment by using the LDFSRSs and SRLDFSs. The intelligent system, which consists of both technical and economical criteria in the material handling equipment selection process, is presented in Figure 2.

4.1. Selection of a Sustainable Material Handling Equipment by Using LDFSRSs

We suppose that a manufacturing company wants to increase efficiency and needs to deal with the materials professionally. The company wants to select that alternative that decreases the lead times and increases productivity. After some basic assessment, the board of the company constructs the set of suitable alternatives given as Q ¨ = { G 1 , G 2 , G 3 , G 4 , G 5 , G 6 , G 7 } . To measure the appropriate alternative, several decision makers from the company’s technical board are organized. They choose some significant decision variables according to their requirements, given as set G ˙ = { ˙ 1 , ˙ 2 , ˙ 3 , ˙ 4 } , where:
˙ 1 = Technical : convenience , maintainability , safety   required , ˙ 2 = Monetary : setting   up   and   oprrational   cost , maintenance   cost , purchasing   cost , ˙ 3 = Operational : fuel   consumption , moving   speed , capacity , ˙ 4 = Strategic : flexibility , level   of   training   required , guarantee .
We divide the attributes into sub-criteria under the effect of parameterizations. This categorizes the data and gives us a wide domain for the selection of truth and falsity grades for the alternatives to the corresponding decision variables. The categorization is given as follows:
  • “Technical: convenience, maintainability, safety required” means that the alternative is “highly technical” or may be “low”.
  • “Monetary: setting up and operational cost, maintenance cost, purchasing cost” means that the alternative may be “expansive” or “inexpensive”.
  • “Operational: fuel consumption, moving speed, capacity” means that the alternative is “highly operational” or may be “low”.
  • “Strategic: flexibility, level of training required, guarantee” means that the alternative is “highly strategic” or may be “low”.
Table 10 represents the sub-attributes of the listed criteria.
We developed two novel algorithms (Algorithms 1 and 2) for the selection of best material handling equipment by using LDFSRSs. The flowchart diagram of both algorithms is given in Figure 3.
Algorithm 1: Selection of a best material handling equipment by using LDFSRSs.
Input:
1. Input the reference set Q ¨ .
2. Input the assembling of attributes G ˙ .
Construction:
3. According to the necessity of the DM, build an LDFSR ð ˜ : Q ¨ G ˙ .
4. Based on the needs of the decision maker, construct LDF-subset B D of G ˙ as an optimal normal decision set.
Calculation:
5. Calculate the “LDFSR approximation operators” ð ˜ ( B D ) and ð ˜ ( B D ) as lower and upper using Definition 12.
6. By using Definition 13 of the ring sum operation, find the choice of LDFS ð ˜ ( B D ) ð ˜ ( B D ) .
Output:
7. We use the definitions of score, quadratic score, and expectation score functions for LDFNs A ¨ D = ( t ˙ D , f ˙ D , α D , β D ) given in [54] and written respectively as:
L 1 ( A ¨ D ) = 1 2 [ ( t ˙ D f ˙ D ) + ( α D β D ) ]

L 2 ( A ¨ D ) = 1 2 [ ( t ˙ D 2 f ˙ D 2 ) + ( α D 2 β D 2 ) ]

L 3 ( A ¨ D ) = 1 2 [ ( t ˙ D f ˙ D + 1 ) 2 + ( α D β D + 1 ) 2 ]
of every alternative in ð ˜ ( B D ) ð ˜ ( B D ) .
8. Rank the alternatives by using calculated score values.
Final decision:
9. Choose the alternative having the maximum score value.
Algorithm 2: Selection of the best material handling equipment by using LDFSRSs.
Input:
1. Input the reference set Q ¨ .
2. Input the assembling of attributes G ˙ .
Construction:
3. According to the necessity of the DM, build an LDFSR ð ˜ : Q ¨ G ˙ .
4. Based on the needs of the decision maker, construct LDF-subset B D of G ˙ as an optimal normal decision set.
Calculation:
5. Calculate the “LDFSR approximation operators” ð ˜ ( B D ) and ð ˜ ( B D ) as lower and upper using Definition 12.
6. For “ N ” number of experts, calculate upper and lower reducts from the calculated “upper and lower approximation operators”, respectively.
Output:
7. From the calculated “ 2 N ” reducts, we get “ 2 N ” crisp subsets of the reference set Q ¨ . The subsets can be constructed by using the “YES” and “NO” logic. The only alternatives in the reduct having final decision “YES” will become the object of the crisp subset.
8. Calculate the core set by taking the intersection of all crisp subsets obtained from the calculated reducts.
Final decision:
9. The alternatives in the core will be our choice for the final decision.

4.1.1. Calculations by Using Algorithm 1

The indiscernibility relation is “the selection of best material handling equipment”. This relation can be observed by LDFSR, ð ˜ : Q ¨ G ˙ given as Table 11.
Thus, ð ˜ is an LDFSR on Q ¨ × G ˙ . This relation gives us the numeric values in the form of LDFNs of each alternative corresponding to every decision variable. For example, for the alternative G 1 , the decision variable ˙ 1 (“Technical: convenience, maintainability, safety required”) has numeric value ( 0.73 , 0.41 , 0.31 , 0.13 ) . This value shows that the alternative G 1 is 73 % technical and 41 % has a falsity value for technicality. The pair 0.31 , 0.13 represents the reference parameters for the truth and falsity grades, where we can observe that alternative G 1 is 31 % highly technical and it has 13 % low technicality. These sub-criteria for the alternatives can be observed from Table 10. All the remaining values can be constructed according to a similar pattern. We consider that experts give some opinion about the attributes and rank them according to their requirement. We convert the verbal description into the LDFS numeric values in the form of LDFS B D . The set B D is the LDF-subset of G ˙ and written as follows:
B D = { ( ˙ 1 , 00.63 , 00.41 , 00.31 , 00.33 ) , ( ˙ 2 , 00.71 , 00.51 , 00.41 , 00.38 ) , ( ˙ 3 , 00.75 , 00.63 , 00.51 , 00.32 ) , ( ˙ 4 , 00.83 , 00.51 , 00.41 , 00.21 ) } .
We evaluate the “lower and upper approximations” of LDFS B D on LDFSR ð ˜ .
ð ˜ ( B D ) = { ( G 1 , 00.73 , 00.51 , 00.31 , 00.64 ) , ( G 2 , 00.71 , 00.57 , 00.51 , 00.58 ) , ( G 3 , 00.71 , 00.51 , 00.41 , 00.59 ) , ( G 4 , 00.75 , 00.41 , 00.41 , 00.58 ) , ( G 5 , 00.83 , 00.51 , 00.41 , 00.58 , ( G 6 , 00.71 , 00.41 , 00.38 , 00.58 , ( G 7 , 00.71 , 00.42 , 00.41 , 00.51 ) } ð ˜ ( B D ) = { ( G 1 , 00.63 , 00.51 , 00.69 , 00.23 ) , ( G 2 , 00.63 , 00.41 , 00.41 , 00.33 ) , ( G 3 , 00.63 , 00.51 , 00.49 , 00.38 ) , ( G 4 , 00.63 , 00.51 , 00.39 , 00.38 ) , ( G 5 , 00.63 , 00.49 , 00.49 , 00.32 , ( G 6 , 00.63 , 00.49 , 00.59 , 00.38 , ( G 7 , 00.63 , 00.63 , 00.51 , 00.38 ) }
ð ˜ ( B D ) ð ˜ ( B D ) = { ( G 1 , 0.900 , 0.260 , 0.780 , 0.140 ) , ( G 2 , 0.890 , 0.230 , 0.710 , 0.190 ) , ( G 3 , 0.890 , 0.260 , 0.690 , 0.220 ) , ( G 4 , 0.900 , 0.200 , 0.640 , 0.220 ) , ( G 5 , 0.937 , 0.249 , 0.699 , 0.185 , ( G 6 , 0.892 , 0.200 , 0.745 , 0.220 , ( G 7 , 0.890 , 0.260 , 0.710 , 0.190 ) }
Now, we calculate the score values, quadratic score values, and expectation score values of the alternatives in ð ˜ ( B D ) ð ˜ ( B D ) . The final ranking is given in Table 12.
From Table 12, we can observe that the alternative G 1 is most suitable for the final decision. The bar chart of the ranking results for alternatives is given in Figure 4.

4.1.2. Calculations by Using Algorithm 2

In Algorithm 1, we use the input data in the form of linguistic terms as LDFNs. We only deal with the truth and falsity grades with their reference parameters, and we have no idea about the expert’s opinion. Due to the lack of information, we have some uncertainty in our decision. This uncertainty can be removed by giving some weight to the expert’s opinion. Therefore, we establish upper and lower reducts for all the experts one by one. The initial five steps of Algorithm 2 are the same as Algorithm 1. We will proceed next by constructing the upper and lower reducts from “upper and lower approximations” of LDFS for all the experts. Suppose that we have three experts from the company’s technical committee given as:
Expert   X Expert   Y Expert   Z
The reducts from approximations can be constructed by using the following terms.
T ¨ D = Truth   membership   grade , S ¨ D = Falsity   membership   grade , α D = Reference   parameter   corresponding   to   truth   membership   grade , β D = Reference   parameter   corresponding   to falsity   membership   grade , L 3 = Expectation   score   function   value   of   LDFN , L ^ = Ranking   given   by   experts   to   the   alternatives   from   crisp   set   { 0 , 1 } , L * = Selection   of   alternative   by   using   YES or NO , i . e . , take   average   of scores   L 3 for   all   the   alternatives . The   alternatives   that   have   a   grater or   equal   score   L 3 than / to   the   average   can   be   seleted   as YES ; those   who have   a lesser   score   than   the   average   value   can   be   neglected   as   NO , F . D = Fianl   decision
The final decision is based on L ^ and L * given in Table 13.
For expert-X, the upper reduct of upper approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 14. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.520 .
This implies that U X = { G 5 , G 6 , G 7 } . For expert-X, the lower reduct of lower approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 15. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.572 .
This implies that L X = { G 1 , G 5 , G 6 } . For expert-Y, the upper reduct of upper approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 16. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.520 .
This implies that U Y = { G 4 , G 5 , G 6 } . For expert-Y, the lower reduct of lower approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 17. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.572 .
This implies that L Y = { G 2 , G 5 , G 6 } . For expert-Z, the upper reduct of upper approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 18. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.520 .
This implies that U Z = { G 5 , G 7 } . For expert-Z, the lower reduct of lower approximation ð ˜ ( B D ) (calculated in Algorithm 1) of LDFS B D is given as Table 19. The average of the score values of all the alternatives for ð ˜ ( B D ) is 0.572 .
This implies that L Z = { G 1 , G 2 , G 5 } . Now, we calculate the core set by taking the intersection of all upper and lower reducts for all three experts.
core = U X L X U Y L Y U Z L Z = { G 5 }
This means that “ G 5 ” is the most suitable alternative for the final decision.

4.2. Selection of the Most Appropriate Material Handling Equipment by Using SRLDFSs

Now, we use our second novel structure of SRLDFS and “crisp soft approximation space” for the selection of the most appropriate material handling equipment. We construct two novel algorithms (Algorithms 3 and 4) for the selection. The flowchart diagram of both algorithms is given in Figure 5.
Algorithm 3: Selection of the best material handling equipment by using SRLDFSs.
Input:
1. Input the reference set Q ¨ .
2. Input the assembling of attributes G ˙ .
Construction:
3. According to the necessity of the DM, build a crisp soft relation A ˜ over Q ¨ × G ˙ .
4. Based on the needs of the decision maker, construct LDF-subset H of G ˙ as an optimal normal decision set.
Calculation:
5. Calculate the “SRLDF approximation operators” A ˜ ( H ) and A ˜ ( H ) as “lower and upper approximations” by using Definition 9.
6. By using Definition 13 of the ring sum operation, find the choice of LDFS A ˜ ( H ) A ˜ ( H ) .
Output:
7. We use the definitions of the score, quadratic score, and expectation score functions for LDFNs A ¨ D = ( t ˙ D , f ˙ D , α D , β D ) given in [54] and written respectively as:
L 1 ( A ¨ D ) = 1 2 [ ( t ˙ D f ˙ D ) + ( α D β D ) ]

L 2 ( A ¨ D ) = 1 2 [ ( t ˙ D 2 f ˙ D 2 ) + ( α D 2 β D 2 ) ]

L 3 ( A ¨ D ) = 1 2 [ ( t ˙ D f ˙ D + 1 ) 2 + ( α D β D + 1 ) 2 ]
of every alternative in A ˜ ( H ) A ˜ ( H ) .
8. Rank the alternatives by using calculated score values.
Final decision:
9. Select the object having the highest score value.
Algorithm 4: Selection of the best material handling equipment by using SRLDFSs.
Input:
1. Input the reference set Q ¨ .
2. Input the assembling of attributes G ˙ .
Construction:
3. According to the necessity of the DM, build a crisp soft relation A ˜ over Q ¨ × G ˙ .
4. Based on the needs of the decision maker, construct LDF-subset H of G ˙ as an optimal normal decision set.
Calculation:
5. Calculate the “SRLDF approximation operators” A ˜ ( H ) and A ˜ ( H ) as “lower and upper approximations” by using Definition 9.
6. For “ N ” number of experts, calculate upper and lower reducts from the calculated “upper and lower approximation operators”, respectively.
Output:
7. From calculated “ 2 N ” reducts, we get “ 2 N ” crisp subsets of the reference set Q ¨ . The subsets can be constructed by using the “YES” and “NO” logic. The only alternatives in the reduct having final decision “YES” will become the object of the crisp subset.
8. Calculate the core set by taking the intersection of all crisp subsets obtained from the calculated reducts.
Final decision:
9. The alternatives in the core will be our choice for the final decision.

4.2.1. Calculations by Using Algorithm 3

We consider the indiscernibility relation “selection of best material handling equipment”. This relation is represented as a crisp soft relation A ˜ over Q ¨ × G ˙ given as Table 20.
Thus, A ˜ over Q ¨ × G ˙ is a crisp soft relation. Table 20 shows that we have:
A ˜ s ( G 1 ) = { ˙ 3 , ˙ 4 } A ˜ s ( G 2 ) = { ˙ 2 , ˙ 3 } A ˜ s ( G 3 ) = { ˙ 1 , ˙ 2 } A ˜ s ( G 4 ) = { ˙ 1 , ˙ 2 , ˙ 3 } A ˜ s ( G 5 ) = { ˙ 1 , ˙ 2 , ˙ 4 } A ˜ s ( G 6 ) = { ˙ 1 , ˙ 4 } A ˜ s ( G 7 ) = { ˙ 2 , ˙ 3 }
We consider that experts give some opinion about the attributes and rank them according to their requirements. We convert the verbal description into the LDFS numeric values in the form of LDFS H . The set H is the LDF-subset of G ˙ and written as follows:
H = { ( ˙ 1 , 0.63 , 0.41 , 0.31 , 0.33 ) , ( ˙ 2 , 0.71 , 0.51 , 0.41 , 0.38 ) , ( ˙ 3 , 0.75 , 0.63 , 0.51 , 0.32 ) , ( ˙ 4 , 0.83 , 0.51 , 0.41 , 0.21 ) } .
Now, we find “upper and lower approximations” of set H over the relation A ˜ by using Definition 9 given as:
A ˜ ( H ) = { ( G 1 , 00.83 , 00.51 , 00.51 , 00.21 ) , ( G 2 , 00.75 , 00.51 , 00.51 , 00.32 ) , ( G 3 , 00.71 , 00.41 , 00.41 , 00.33 ) , ( G 4 , 00.75 , 00.41 , 00.51 , 00.32 ) , ( G 5 , 00.83 , 00.41 , 00.41 , 00.21 , ( G 6 , 00.83 , 00.41 , 00.41 , 00.21 , ( G 7 , 00.75 , 00.51 , 00.51 , 00.32 ) } A ˜ ( H ) = { ( G 1 , 00.75 , 00.63 , 00.41 , 00.32 ) , ( G 2 , 00.71 , 00.63 , 00.41 , 00.38 ) , ( G 3 , 00.63 , 00.51 , 00.31 , 00.38 ) , ( G 4 , 00.63 , 00.63 , 00.31 , 00.38 ) , ( G 5 , 00.63 , 00.51 , 00.31 , 00.38 , ( G 6 , 00.63 , 00.51 , 00.31 , 00.33 , ( G 7 , 00.71 , 00.63 , 00.41 , 00.38 ) }
A ˜ ( H ) A ˜ ( H ) = { ( G 1 , 0.957 , 0.321 , 0.710 , 0.067 ) , ( G 2 , 0.927 , 0.321 , 0.710 , 0.121 ) , ( G 3 , 0.892 , 0.209 , 0.592 , 0.125 ) , ( G 4 , 0.907 , 0.258 , 0.661 , 0.121 ) , ( G 5 , 0.937 , 0.209 , 0.592 , 0.079 , ( G 6 , 0.937 , 0.209 , 0.592 , 0.069 , ( G 7 , 0.927 , 0.321 , 0.710 , 0.121 ) }
Now, we calculate the score values, quadratic score values, and expectation score values of alternatives in A ˜ ( H ) A ˜ ( H ) . The calculated data with the final ranking is given in Table 21.
From Table 21, we can observe that the alternative G 1 is most suitable for the final decision. The bar chart of the ranking results for alternatives is given in Figure 6.

4.2.2. Calculations by Using Algorithm 4

In this part, we establish upper and lower reducts for all the experts one by one. The initial five steps of Algorithm 4 are the same as Algorithm 3. We will proceed next by constructing the upper and lower reducts from the “upper and lower approximations” of LDFS for all the experts under “crisp soft approximation space”. Suppose that we have three experts from the company’s technical committee given as:
Expert   X ˙ Expert   Y ˙ Expert   Z ˙
The characteristics and terms for finding the upper and lower reducts are the same as we used in Algorithm 2. Therefore, we directly calculate the reducts for experts.
For expert- X ˙ , the upper reduct of upper approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 22. The average of the score values of all the alternatives for A ˜ ( H ) is 0.629 .
This implies that ( U X ˙ ) = { G 1 , G 5 , G 6 } . For expert- X ˙ , the lower reduct of lower approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 23. The average of the score values of all the alternatives for A ˜ ( H ) is 0.519 .
This implies that L X ˙ = { G 1 , G 2 , G 6 } . For expert- Y ˙ , the upper reduct of upper approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 24. The average of the score values of all the alternatives for A ˜ ( H ) is 0.629 .
This implies that ( U Y ˙ ) = { G 1 , G 4 , G 5 } . For expert- Y ˙ , the lower reduct of lower approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 25. The average of the score values of all the alternatives for A ˜ ( H ) is 0.519 .
This implies that L Y ˙ = { G 1 , G 7 } . For expert- Z ˙ , the upper reduct of upper approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 26. The average of the score values of all the alternatives for A ˜ ( H ) is 0.629 .
This implies that ( U Z ˙ ) = { G 1 , G 4 , G 6 } . For expert- Z ˙ , the lower reduct of lower approximation A ˜ ( H ) (calculated in Algorithm 3) of LDFS H is given as Table 27. The average of the score values of all the alternatives for A ˜ ( H ) is 0.519 .
This implies that L Z ˙ = { G 1 , G 2 , G 6 , G 7 } .
Now, we calculate the core set by taking the intersection of all upper and lower reducts for all three experts.
core = U X ˙ L X ˙ U Y ˙ L Y ˙ U Z ˙ L Z ˙ = { G 1 }
This means that “ G 1 ” is the most suitable alternative for the final decision.

4.3. Discussion, Comparison, and Symmetrical Analysis

In this part, we compare our models to the existing approaches and discuss the superiority, authenticity, symmetry, and validity of our proposed structures. The comparison of the proposed structures with existing models is shown in Table 28 and Table 29. Such tables reflect the characteristics and limitations of certain current hypotheses. We will observe that our presented models are superior and handle the MCDM techniques efficiently.
We constructed four algorithms based on LDFSRSs, SRLDFSs, and their corresponding approximation spaces. The final results for the decision making problem of material handling equipment selection obtained from these algorithms are given in Table 30.
In existing work, the superiority of the proposed model was discussed by examining its degeneration towards some existing rough set models (see Table 5 and Table 9). The proposed algorithms are based on the SRLDFSs and LDFSRSs and their approximation operators. Algorithms 1 and 3 are based on the structures with LDFN score values. These algorithms provide us with information about the best and worst alternative. Algorithms 2 and 4 are focused on the core and reducts of the suggested structures. This also involves expert opinion and produces an outcome only for the essential alternative. This does not offer any comparison of the alternatives. Depending on the situation, each algorithm is essential and useful for real-life issues (see Table 12 and Table 21).
By using different score functions and evaluating the reducts and core set, we check the behavior of “upper and lower approximations”. The final results of Algorithms 1, 3, and 4 are exactly the same. The result of Algorithm 2 is different from the others. This difference is due to the different formulae and different ordering strategies used in the proposed algorithms. As we can see, the three algorithms produce the same decision, so we will go with the alternative G 1 for the final decision. Such structures demonstrate the symmetry in the findings and provide us with an appropriate, ideal approach for the problem of decision making.
Validity test:
To demonstrate the validity and symmetry of the results, Wang and Triantaphyllou [36] constructed the following test criteria.
Test Criterion 1:
“If we replace non-optimal alternative rating values with the worst alternative then the best alternative should not change, provided the relative weighted criteria remain unchanged”.
Test Criterion 2:
“Process should have transitive nature”.
Test Criterion 3:
“When a given problem is decomposed into smaller ones and the same MCDM method has been applied, then the combined ranking of alternatives should be identical to the ranking of un-decomposed one”.
Via these parameters, when we test our results, we see that our findings are correct and reliable and provide us a satisfactory solution to the MCDM problem. Various researchers used numerous techniques based on rough set theory and its hybrid structures to solve decision making difficulties (see [2,3,7,8,17,18,19,23,24,26,33,34,35]). Comparing these hypotheses, we found that our proposed models are reliable, efficient, superior, symmetrical, and valid in comparison with those current models.

5. Conclusions

There are two viewpoints in rough set theory knowledge: positive and axiomatic methods, and it is the same for LDFSRSs and SRLDFSs. This manuscript is a crystal reflection of both aspects of it. We have practiced fundamental ingredients of rough sets, soft sets, and LDFSs and established the proposed structures. With their accompanying illustrations, we provided some findings of such models. Many of the barriers to decision making in the input dataset include unclear, ambiguous, and imprecise details. These models can control these ambiguities better than the fuzzy sets, IFSs, PFSs, q-ROFSs, and LDFSs due to their mathematical formulation, variations, symmetry, and novelty. We introduced several level cut sets of LDFSs and related the recommended approximation operators with these level cut relations. We established various illustrations and results based on LDFSRSs and SRLDFSs approximation operators and corresponding approximations based on level cut sets. We utilized two different approximation spaces to produce variety in the decision making results. We listed the results of the degeneration of the proposed operators and found that our proposed models are generalizations of various existing rough set models. By using approximation spaces, score functions, upper and lower reductions, and core series, we introduced four novel algorithms for the assortment of sustainable material handling equipment. Depending on the situation, each algorithm is essential and useful for solving real-life problems. We discussed the advantages and limitations of the proposed structures with some existing models briefly (see Table 1). In the future, we will expand this research for topological spaces and solve MCDM problems based on the TOPSIS, VIKOR, and AHP families.

Author Contributions

M.R., M.R.H., H.K., D.P., and Y.-M.C. conceived of and worked together to achieve this manuscript; M.R., M.R.H., and D.P. constructed the ideas and algorithms for data analysis and designed the model of the manuscript; M.R.H., H.K., and Y.-M.C. processed the data collection and wrote the paper. All authors read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors are highly thankful to the Editor-in-Chief and referees for their valuable comments and suggestions for the improvement of our manuscript.

Conflicts of Interest

The authors declare that they have no conflict of interest.

Abbreviations

FSsFuzzy sets
IFSsIntuitionistic fuzzy sets
PFSsPythagorean fuzzy sets
q-ROFSsq-rung orthopair fuzzy sets
LDFSsLinear Diophantine fuzzy sets
LDFNsLinear Diophantine fuzzy numbers
LDFSSsLinear Diophantine fuzzy soft sets
LDFSRSsLinear Diophantine fuzzy soft rough sets
SRLDFSsSoft rough linear Diophantine fuzzy sets
MCDMMulti-criteria decision making

Appendix A

(1)
From Definition 9, we can write that:
A ˜ ( Y D ) = { ( G , S ¨ A ˜ ( Y D ) ( G ) , T ¨ A ˜ ( Y D ) ( G ) , β A ˜ ( Y D ) ( G ) , α A ˜ ( Y D ) ( G ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) ( S ¨ ( Y D ) ( ˙ ) ) , max ˙ A ˜ s ( G ) ( T ¨ ( Y D ) ( ˙ ) ) , min ˙ A ˜ s ( G ) ( β ( Y D ) ( ˙ ) ) , max ˙ A ˜ s ( G ) ( α ( Y D ) ( ˙ ) ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) ( T ¨ Y D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( S ¨ Y D ( ˙ ) ) , min ˙ A ˜ s ( G ) ( α Y D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( β Y D ( ˙ ) ) ) : G Q ¨ } = { ( G , T ¨ A ˜ ( Y D ) ( G ) , S ¨ A ˜ ( Y D ) ( G ) , α A ˜ ( Y D ) ( G ) , β A ˜ ( Y D ) ( G ) ) : G Q ¨ } = A ˜ ( Y D )
(2)
It can be easily proven from Definition 9.
(3)
We consider that:
A ˜ ( Y D B D ) = { ( G , T ¨ A ˜ ( Y D B D ) ( G ) , S ¨ A ˜ ( Y D B D ) ( G ) , α A ˜ ( Y D B D ) ( G ) , β A ˜ ( Y D B D ) ( G ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) T ¨ ( Y D B D ) ( ˙ ) , max ˙ A ˜ s ( G ) S ¨ ( Y D B D ) ( ˙ ) , min ˙ A ˜ s ( G ) α ( Y D B D ) ( ˙ ) , max ˙ A ˜ s ( G ) β ( Y D B D ) ( ˙ ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) ( T ¨ Y D ( ˙ ) α T ¨ B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( S ¨ Y D ( ˙ ) α S ¨ B D ( ˙ ) ) , min ˙ A ˜ s ( G ) ( α Y D ( ˙ ) α T ¨ B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( β Y D ( ˙ ) α S ¨ B D ( ˙ ) ) ) } = { ( G , min ˙ A ˜ s ( G ) ( T ¨ Y D ( ˙ ) ) min ˙ A ˜ s ( G ) ( T ¨ B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( S ¨ Y D ( ˙ ) ) max ˙ A ˜ s ( G ) ( S ¨ B D ( ˙ ) ) , min ˙ A ˜ s ( G ) ( α Y D ( ˙ ) ) min ˙ A ˜ s ( G ) ( α B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( β Y D ( ˙ ) ) max ˙ A ˜ s ( G ) ( β B D ( ˙ ) ) ) } = { ( G , T ¨ A ˜ ( Y D ) ( G ) T ¨ A ˜ ( B D ) ( G ) , S ¨ A ˜ ( Y D ) ( G ) S ¨ A ˜ ( B D ) ( G ) , α A ˜ ( Y D ) ( G ) α A ˜ ( B D ) ( G ) , β A ˜ ( Y D ) ( G ) β A ˜ ( B D ) ( G ) ) : G Q ¨ } = A ˜ ( Y D ) A ˜ ( B D )
(4)
From Definition 9, we can write that:
A ˜ ( Y D B D ) = { ( G , T ¨ A ˜ ( Y D B D ) ( G ) , S ¨ A ˜ ( Y D B D ) ( G ) , α A ˜ ( Y D B D ) ( G ) , β A ˜ ( Y D B D ) ( G ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) T ¨ ( Y D B D ) ( ˙ ) , max ˙ A ˜ s ( G ) S ¨ ( Y D B D ) ( ˙ ) , min ˙ A ˜ s ( G ) α ( Y D B D ) ( ˙ ) , max ˙ A ˜ s ( G ) β ( Y D B D ) ( ˙ ) ) : G Q ¨ } = { ( G , min ˙ A ˜ s ( G ) ( T ¨ Y D ( ˙ ) α T ¨ B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( S ¨ Y D ( ˙ ) S ¨ B D ( ˙ ) ) , min ˙ A ˜ s ( G ) ( α Y D ( ˙ ) α T ¨ B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( β Y D ( ˙ ) S ¨ B D ( ˙ ) ) ) } { ( G , min ˙ A ˜ s ( G ) ( T ¨ Y D ( ˙ ) ) min ˙ A ˜ s ( G ) ( T ¨ B D ( ˙ ) ) , ˙ A ˜ s ( G ) ( S ¨ Y D ( ˙ ) ) max ˙ A ˜ s ( G ) ( S ¨ B D ( ˙ ) ) , min ˙ A ˜ s ( G ) ( α Y D ( ˙ ) ) min ˙ A ˜ s ( G ) ( α B D ( ˙ ) ) , max ˙ A ˜ s ( G ) ( β Y D ( ˙ ) ) max ˙ A ˜ s ( G ) ( β B D ( ˙ ) ) ) } = { ( G , T ¨ A ˜ ( Y D ) ( G ) T ¨ A ˜ ( B D ) ( G ) , S ¨ A ˜ ( Y D ) ( G ) S ¨ A ˜ ( B D ) ( G ) , α A ˜ ( Y D ) ( G ) α A ˜ ( B D ) ( G ) , β A ˜ ( Y D ) ( G ) β A ˜ ( B D ) ( G ) ) : G Q ¨ } = A ˜ ( Y D ) A ˜ ( B D )
Thus, A ˜ ( Y D B D ) A ˜ ( Y D ) A ˜ ( B D ) .
Similarly, we can prove the remaining axioms by following these arguments.

References

  1. Bellman, R.E.; Zadeh, L.A. Decision-making in a fuzzy environment. Manag. Sci. 1970, 4, 141–164. [Google Scholar] [CrossRef]
  2. Akram, M.; Ali, G.; Alshehri, N.O. A new multi-attribute decision making method based on m-polar fuzzy soft rough sets. Symmetry 2017, 9, 271. [Google Scholar] [CrossRef] [Green Version]
  3. Ali, M.I. A note on soft sets, rough soft sets and fuzzy soft sets. Appl. Soft Comput. 2011, 11, 3329–3332. [Google Scholar]
  4. Chen, S.M.; Tan, J.M. Handling multi-criteria fuzzy decision-makling problems based on vague set theory. Fuzzy Sets Syst. 1994, 67, 163–172. [Google Scholar] [CrossRef]
  5. Tversky, A.; Kahneman, D. Advances in prospect theory: Cumulative representation of uncertainity. J. Risk Uncertain. 1992, 5, 297–323. [Google Scholar] [CrossRef]
  6. Feng, F.; Jun, Y.B.; Liu, X.; Li, L. An adjustable approach to fuzzy soft set based decision making. J. Comput. Appl. Math. 2010, 234, 10–20. [Google Scholar] [CrossRef] [Green Version]
  7. Feng, F.; Li, C.; Davvaz, B.; Ali, M.I. Soft sets combined with fuzzy sets and rough sets; A tentative approach. Soft Comput. 2010, 14, 899–911. [Google Scholar] [CrossRef]
  8. Feng, F.; Liu, X.Y.; Leoreanu-Fotea, V.; Jun, Y.B. Soft sets and soft rough sets. Inf. Sci. 2011, 181, 1125–1137. [Google Scholar] [CrossRef]
  9. Feng, F.; Fujita, H.; Ali, M.I.; Yager, R.R.; Liu, X. Another view on generalized intuitionistic fuzzy soft sets and related multiattribute decision making methods. IEEE Trans. Fuzzy Syst. 2019, 27, 474–488. [Google Scholar] [CrossRef]
  10. Garg, H. A new generalized Pythagorean fuzzy information aggregation using Einstein operations and its application to decision making. Int. J. Intell. Syst. 2016, 31, 886–920. [Google Scholar] [CrossRef]
  11. Hashmi, M.R.; Riaz, M.; Smarandache, F. m-polar Neutrosophic Topology with Applications to Multi-Criteria Decision-Making in Medical Diagnosis and Clustering Analysis. Int. J. Fuzzy Syst. 2020, 22, 273–292. [Google Scholar] [CrossRef]
  12. Hashmi, M.R.; Riaz, M. A Novel Approach to Censuses Process by using Pythagorean m-polar Fuzzy Dombi’s Aggregation Operators. J. Intell. Fuzzy Syst. 2020, 38, 1977–1995. [Google Scholar] [CrossRef]
  13. Jose, S.; Kuriaskose, S. Aggregation operators, score function and accuracy function for multi criteria decision making in intuitionistic fuzzy context. Notes Intuitionistic Fuzzy Sets 2014, 20, 40–44. [Google Scholar]
  14. Naeem, K.; Riaz, M.; Peng, X.; Afzal, D. Pythagorean fuzzy soft MCGDM methods based on TOPSIS, VIKOR and aggregation operators. J. Intell. Fuzzy Syst. 2019, 37, 6937–6957. [Google Scholar] [CrossRef]
  15. Pawlak, Z.; Skowron, A. Rough sets: Some extensions. Inf. Sci. 2007, 177, 28–40. [Google Scholar] [CrossRef] [Green Version]
  16. Riaz, M.; Hashmi, M.R. MAGDM for agribusiness in the environment of various cubic m-polar fuzzy averaging aggregation operators. J. Intell. Fuzzy Syst. 2019, 37, 3671–3691. [Google Scholar] [CrossRef]
  17. Riaz, M.; Hashmi, M.R. Soft Rough Pythagorean m-Polar Fuzzy Sets and Pythagorean m-Polar Fuzzy Soft Rough Sets with Application to Decision-Making. Comput. Appl. Math. 2020, 39, 16. [Google Scholar] [CrossRef]
  18. Riaz, M.; Davvaz, B.; Firdous, A.; Fakhar, A. Novel Concepts of Soft Rough Set Topology with Applications. J. Intell. Fuzzy Syst. 2019, 36, 3579–3590. [Google Scholar] [CrossRef]
  19. Riaz, M.; Samrandache, F.; Firdous, A.; Fakhar, A. On Soft Rough Topology with Multi-Attribute Group Decision Making. Mathematics 2019, 7, 67. [Google Scholar] [CrossRef] [Green Version]
  20. Riaz, M.; Tehrim, S.T. Certain properties of bipolar fuzzy soft topology via Q-neighborhood. Punjab Univ. J. Math. 2019, 51, 113–131. [Google Scholar]
  21. Riaz, M.; Tehrim, S.T. Cubic bipolar fuzzy ordered weighted geometric aggregation operators and their application using internal and external cubic bipolar fuzzy data. Comput. Appl. Math. 2019, 38, 87. [Google Scholar] [CrossRef]
  22. Riaz, M.; Tehrim, S.T. Multi-attribute group decision making based cubic bipolar fuzzy information using averaging aggregation operators. J. Intell. Fuzzy Syst. 2019, 37, 2473–2494. [Google Scholar] [CrossRef]
  23. Roy, J.; Adhikary, K.; Kar, S.; Pamucar, D. A rough strength relational DEMATEL model for analysing the key success factors ofhospital service quality. Decis. Mak. Appl. Manag. Eng. 2018, 1, 121–142. [Google Scholar] [CrossRef]
  24. Sharma, H.K.; Kumari, K.; Kar, S. A rough set theory application in forecasting models. Decis. Mak. Appl. Manag. Eng. 2020, 3, 1–21. [Google Scholar] [CrossRef]
  25. Wei, G.; Wang, H.; Zhao, X.; Lin, R. Hesitant triangular fuzzy information aggregation in multiple attribute decision making. J. Intell. Fuzzy Syst. 2014, 26, 1201–1209. [Google Scholar] [CrossRef]
  26. Zhang, H.; Shu, L.; Liao, S. Intuitionistic fuzzy soft rough set and its applications in decision making. Abstr. Appl. Anal. 2014, 1–13. [Google Scholar] [CrossRef] [Green Version]
  27. Zhao, H.; Xu, Z.S.; Ni, M.F.; Lui, S.S. Generalized aggregation operators for intuitionistic fuzzy sets. Int. J. Intell. Syst. 2010, 25, 1–30. [Google Scholar] [CrossRef]
  28. Xu, Z.S.; Chen, J. An overview on distance and similarity measures on intuitionistic fuzzy sets. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 2008, 16, 529–555. [Google Scholar] [CrossRef]
  29. Kulak, O.; Satoglu, S.I.; Durmusoglu, M.B. Multi-attribute material handling equipment selection using information axiom. In Proceedings of the ICAD2004, The third international conferenec on Axiomatic Design, Seoul, Korea, 21–24 June 2004; pp. 1–9. [Google Scholar]
  30. Karande, P.; Chakraborty, S. Material handling equipment selection using weighted utility additive theory. J. Ind. Eng. 2013. [Google Scholar] [CrossRef]
  31. Zubair, M.; Mahmood, S.; Omair, M.; Noor, I. Optimization of material handling system through material handling equipment selection. Int. J. Progress. Sci. Technol. 2019, 15, 235–243. [Google Scholar]
  32. Vashist, R. An algorithm for finding the reduct and core of the consistent dataset. Int. Conf. Comput. Intell. Commun. Netw. 2015. [Google Scholar] [CrossRef]
  33. Zhang, L.; Zhan, J. Fuzzy soft β-covering based fuzzy rough sets and corresponding decision making applications. Int. J. Mach. Learn. Cybern. 2018. [Google Scholar] [CrossRef]
  34. Zhang, L.; Zhan, J. Novel classes of fuzzy soft;-coverings-based fuzzy rough sets with applications to multi-criteria fuzzy group decision making. Soft Comput. 2018. [Google Scholar] [CrossRef]
  35. Zhang, L.; Zhan, J.; Xu, Z.X. Covering-based generalized IF rough sets with applications to multi-attribute decision making. Inf. Sci. 2019, 478, 275–302. [Google Scholar] [CrossRef]
  36. Wang, X.; Triantaphyllou, E. Ranking irregularities when evaluating alternatives by using some ELECTRE methods. Omega Int. J. Manag. Sci. 2008, 36, 45–63. [Google Scholar] [CrossRef]
  37. Büyüközkan, G.; Çifçi, G. A novel hybrid MCDM approach based on fuzzy DEMATEL, fuzzy ANP and fuzzy TOPSIS to evaluate green suppliers. Expert Syst. Appl. 2012, 39, 3000–3011. [Google Scholar] [CrossRef]
  38. Govindan, K.; Khodaverdi, R.; Vafadarnikjoo, A. Intuitionistic fuzzy based DEMATEL method for developing green practices and performances in a green supply chain. Expert Syst. Appl. 2015, 42, 7207–7720. [Google Scholar] [CrossRef]
  39. Jeong, J.S.; González-Gómez, D. Adapting to PSTs Pedagogical Changes in Sustainable Mathematics Education through Flipped E-Learning: Ranking Its Criteria with MCDA/F-DEMATEL. Mathematics 2020, 8, 858. [Google Scholar] [CrossRef]
  40. Wang, R.; Li, Y. A Novel Approach for Green Supplier Selection under a q-Rung Orthopair Fuzzy Environment. Symmetry 2018, 10, 687. [Google Scholar] [CrossRef] [Green Version]
  41. Xu, Y.; Shang, X.; Wang, J.; Wu, W.; Huang, H. Some q-Rung Dual Hesitant Fuzzy Heronian Mean Operators with Their Application to Multiple Attribute Group Decision-Making. Symmetry 2018, 10, 472. [Google Scholar] [CrossRef] [Green Version]
  42. Sun, B.; Ma, W. Soft fuzzy rough sets and its application in decision making. Artif. Intell. Rev. 2014, 41, 67–80. [Google Scholar]
  43. Meng, D.; Zhang, X.; Qin, K. Soft rough fuzzy sets and soft fuzzy rough sets. Comput. Math. Appl. 2011, 62, 4635–4645. [Google Scholar] [CrossRef] [Green Version]
  44. Hussain, A.; Ali, M.I.; Mahmood, T. Pythagorean fuzzy soft rough sets and their applications in decision making. J. Taibah Univ. Sci. 2020, 14, 101–113. [Google Scholar] [CrossRef] [Green Version]
  45. Zadeh, L.A. The concept of a linguistic variable and its application to approximate reasoning-I. Inf. Sci. 1975, 8, 199–249. [Google Scholar] [CrossRef]
  46. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef] [Green Version]
  47. Atanassov, K.T. Intuitionistic fuzzy sets. Fuzzy Sets Ans Syst. 1986, 20, 87–96. [Google Scholar] [CrossRef]
  48. Atanassov, K.T. Geometrical interpretation of the elements of the intuitionistic fuzzy objects. Int. J. Bioautom. 2016, 20, S27–S42. [Google Scholar]
  49. Yager, R.R. Pythagorean membership grades in multi-criteria decision making. IEEE Trans. Fuzzy Syst. 2014, 22, 958–965. [Google Scholar] [CrossRef]
  50. Yager, R.R. Pythagorean fuzzy subsets. In Proceedings of the 2013 joint IFSA world congress and NAFIPS annual meeting (IFSA/NAFIPS), Edmonton, AB, Canada, 24–28 June 2013; pp. 57–61. [Google Scholar]
  51. Yager, R.R.; Abbasov, A.M. Pythagorean membership grades, complex numbers, and decision making. Int. J. Intell. Syst. 2013, 28, 436–452. [Google Scholar] [CrossRef]
  52. Ali, M.I. Another view on q-rung orthopair fuzzy sets. Int. J. Intell. Syst. 2018, 33, 2139–2153. [Google Scholar] [CrossRef]
  53. Yager, R.R. Generalized orthopair fuzzy sets. IEEE Trans. Fuzzy Syst. 2017, 25, 1222–1230. [Google Scholar] [CrossRef]
  54. Riaz, M.; Hashmi, M.R. Linear Diophantine fuzzy set and its applications towards multi-attribute decision making problems. J. Intell. Fuzzy Syst. 2019, 37, 5417–5439. [Google Scholar] [CrossRef]
  55. Pawlak, Z. Rough sets. Int. J. Comput. Inf. Sci. 1982, 2, 145–172. [Google Scholar] [CrossRef]
  56. Molodtsov, D. Soft set theory-first results. Comput. Math. Appl. 1999, 37, 19–31. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Graphical comparison among IFNs, PFFNs, q-ROFNs, and LDFNs.
Figure 1. Graphical comparison among IFNs, PFFNs, q-ROFNs, and LDFNs.
Symmetry 12 01215 g001
Figure 2. Configuration of modules in the material handling equipment selection process.
Figure 2. Configuration of modules in the material handling equipment selection process.
Symmetry 12 01215 g002
Figure 3. Flowchart diagram of Algorithms 1 and 2.
Figure 3. Flowchart diagram of Algorithms 1 and 2.
Symmetry 12 01215 g003
Figure 4. Bar chart of alternatives under LDFSRS for SF L 1 , QSF L 2 , and ESF L 3 .
Figure 4. Bar chart of alternatives under LDFSRS for SF L 1 , QSF L 2 , and ESF L 3 .
Symmetry 12 01215 g004
Figure 5. Flowchart diagram of Algorithms 3 and 4.
Figure 5. Flowchart diagram of Algorithms 3 and 4.
Symmetry 12 01215 g005
Figure 6. Bar chart of alternatives under SRLDFS for SF L 1 , QSF L 2 , and ESF L 3 .
Figure 6. Bar chart of alternatives under SRLDFS for SF L 1 , QSF L 2 , and ESF L 3 .
Symmetry 12 01215 g006
Table 1. Semantic comparison of the suggested technique with some existing models.
Table 1. Semantic comparison of the suggested technique with some existing models.
Set TheoriesAdvantagesSemantic Disadvantages
Fuzzy sets [46]Contribute knowledge aboutDose not give information about the
specific propertyfalsity and roughness of information system
Intuitionistic fuzzy sets [47,48]Detect vagueness withRestricted valuation space and
agree and disagree criteriadoes not deal with roughness
Pythagorean fuzzy sets [49,50,51]Detect vagueness with largerCannot handle the roughness of the
valuation space than IFSsdata and dependency between the grades
q-rung orthopair fuzzy sets [52,53]Increase the valuation space of gradesFor smaller values of q, creates dependency
to deal with real-life situationsin grades and cannot handle roughness
Create independency between the degreesDoes not give information about the
Linear Diophantine fuzzy sets [54]and increase their valuation space underroughness of information data and cannot
the effect of control parameterizationsdeal with multi-valued parameterizations
Rough sets [55]Contain upper and lower approximations ofDoes not characterize the agree and
information dataset to handle roughnessdisagree degrees with parameterizations
Soft sets [56]Produce multi-valued mapping basedDoes not contain fuzziness and
parameterizations under different criteriaroughness in optimization
Linear Diophantine fuzzyProduce multi-valued mapping basedDoes not characterize the roughness
soft sets (proposed)on the LDF value information systemof real-life dataset
Contain upper and lower approximationsDue to the use of LDFS approximation
Linear Diophantine fuzzy softwith LDF degrees under doublespace for evaluations, it contains
rough sets (proposed)parameterizations (soft and reference) andheavy calculations, but easy to handle
collect data without any loss of information
Use crisp soft approximation spaceEasy calculations as compared to
Soft rough linear Diophantineto evaluate upper and lower approximationsLDFSRS, but heavy as compared to
fuzzy sets (proposed)with LDF degrees and collect dataothers (easy to handle)
without any loss of information
Table 2. LDFS for medication.
Table 2. LDFS for medication.
AlternativesLDFNs
G 1 ( 00.963 , 00.472 , 00.731 , 00.121 )
G 2 ( 00862 , 00.576 , 00.631 , 00.222 )
G 3 ( 00.665 , 00.874 , 00.332 , 00.123 )
G 4 ( 00.664 , 00.773 , 00.234 , 00.225 )
G 5 ( 00.867 , 00.776 , 00.435 , 00.225 )
Table 3. Description of the notations used in the whole manuscript.
Table 3. Description of the notations used in the whole manuscript.
NotationExplanation
Q ¨ Universal set
G ˙ Set of decision variables
G Elements of set Q ¨
˙ Elements of set G ˙
A ˜ Lower approximation operator for SRLDFSs
A ˜ Upper approximation operator for SRLDFSs
ð ˜ Lower approximation operator for LDFSRSs
ð ˜ Upper approximation operator for LDFSRSs
T ¨ Satisfaction grade
S ¨ Dissatisfaction grade
D ( G ˙ ) Collection of all LDFSs over G ˙
D ( Q ¨ ) Collection of all LDFSs over Q ¨
Table 4. Formulation of SRLDFSs.
Table 4. Formulation of SRLDFSs.
NotationFormulationNotationFormulation
T ¨ A ˜ ( Y D ) ( G ) max ˙ A ˜ s ( G ) T ¨ Y D ( ˙ ) S ¨ A ˜ ( Y D ) ( G ) min ˙ A ˜ s ( G ) S ¨ Y D ( ˙ )
T ¨ A ˜ ( Y D ) ( G ) min ˙ A ˜ s ( G ) T ¨ Y D ( ˙ ) S ¨ A ˜ ( Y D ) ( G ) max ˙ A ˜ s ( G ) S ¨ Y D ( ˙ )
α A ˜ ( Y D ) ( G ) max ˙ A ˜ s ( G ) α Y D ( ˙ ) β A ˜ ( Y D ) ( G ) min ˙ A ˜ s ( G ) β Y D ( ˙ )
α A ˜ ( Y D ) ( G ) min ˙ A ˜ s ( G ) α Y D ( ˙ ) β A ˜ ( Y D ) ( G ) max ˙ A ˜ s ( G ) β Y D ( ˙ )
Table 5. Degeneration of SRLDF approximation operators into different rough set models.
Table 5. Degeneration of SRLDF approximation operators into different rough set models.
Approximation SpaceSet TheoriesFamily of SetsDegeneration of SRLDFAfter Degeneration of the
Approximation OperatorsConstructed Model
Crisp soft Y D =crisp set P ( G ˙ ) P ( G ˙ ) collection of allYesSoft rough
( Q ¨ , G ˙ , A ˜ ) crisp subsets of G ˙ sets [26]
Crisp soft Y D = { ˙ , T ¨ ( ˙ ) : ˙ G ˙ } F ( G ˙ ) F ( G ˙ ) collection of allYesSoft rough
( Q ¨ , G ˙ , A ˜ ) fuzzy subsets of G ˙ fuzzy sets [42,43]
Crisp soft Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } I ( G ˙ ) I ( G ˙ ) collection of allYesSoft rough intuitionistic
( Q ¨ , G ˙ , A ˜ ) IF-subsets of G ˙ fuzzy sets [26]
Crisp soft Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } P ( G ˙ ) P ( G ˙ ) collection of allYesSoft rough Pythagorean
( Q ¨ , G ˙ , A ˜ ) PF-subsets of G ˙ fuzzy sets [44]
Crisp soft Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } R ( G ˙ ) R ( G ˙ ) collection of allYesSoft rough q-rung
( Q ¨ , G ˙ , A ˜ ) PF-subsets of G ˙ orthopair fuzzy sets
Table 6. LDFSR.
Table 6. LDFSR.
ð ˜ ˙ 1 ... ˙ m
G 1 T ¨ ð ˜ ( G 1 , ˙ 1 ) , S ¨ ð ˜ ( G 1 , ˙ 1 ) , α ð ˜ ( G 1 , ˙ 1 ) , β ð ˜ ( G 1 , ˙ 1 ) ... T ¨ ð ˜ ( G 1 , ˙ m ) , S ¨ ð ˜ ( G 1 , ˙ m ) , α ð ˜ ( G 1 , ˙ m ) , β ð ˜ ( G 1 , ˙ m )
G 2 T ¨ ð ˜ ( G 2 , ˙ 1 ) , S ¨ ð ˜ ( G 2 , ˙ 1 ) , α ð ˜ ( G 2 , ˙ 1 ) , β ð ˜ ( G 2 , ˙ 1 ) ... T ¨ ð ˜ ( G 2 , ˙ m ) , S ¨ ð ˜ ( G 2 , ˙ m ) , α ð ˜ ( G 2 , ˙ m ) , β ð ˜ ( G 2 , ˙ m )
............
G n T ¨ ð ˜ ( G n , ˙ 1 ) , S ¨ ð ˜ ( G n , ˙ 1 ) , α ð ˜ ( G n , ˙ 1 ) , β ð ˜ ( G n , ˙ 1 ) ... T ¨ ð ˜ ( G n , ˙ m ) , S ¨ ð ˜ ( G n , ˙ m ) , α ð ˜ ( G n , ˙ m ) , β ð ˜ ( G n , ˙ m )
Table 7. Formulation of LDFSRSs.
Table 7. Formulation of LDFSRSs.
NotationFormulationNotationFormulation
T ¨ ð ˜ ( Y D ) ( G ) max ˙ G ˙ [ T ¨ ð ˜ ( G , ˙ ) T ¨ Y D ( ˙ ) ] S ¨ ð ˜ ( Y D ) ( G ) min ˙ G ˙ [ ( 1 S ¨ ð ˜ ( G , ˙ ) ) S ¨ Y D ( ˙ ) ]
α ð ˜ ( Y D ) ( G ) max ˙ G ˙ [ α ð ˜ ( G , ˙ ) α Y D ( ˙ ) ] β ð ˜ ( Y D ) ( G ) min ˙ G ˙ [ ( 1 β ð ˜ ( G , ˙ ) ) β Y D ( ˙ ) ]
T ¨ ð ˜ ( Y D ) ( G ) min ˙ G ˙ [ ( 1 T ¨ ð ˜ ( G , ˙ ) ) T ¨ Y D ( ˙ ) ] S ¨ ð ˜ ( Y D ) ( G ) max ˙ G ˙ [ S ¨ ð ˜ ( G , ˙ ) S ¨ Y D ( ˙ ) ]
α ð ˜ ( Y D ) ( G ) min ˙ G ˙ [ ( 1 α ð ˜ ( G , ˙ ) ) α Y D ( ˙ ) ] β ð ˜ ( Y D ) ( G ) max ˙ G ˙ [ β ð ˜ ( G , ˙ ) β Y D ( ˙ ) ]
Table 8. LDFSR.
Table 8. LDFSR.
ð ˜ Numeric Values of LDFNs
G 1 ˙ 1 : ( 0.684 , 0.355 , 0.221 , 0.325 )
˙ 2 : ( 0.825 , 0.836 , 0.226 , 0.123 )
˙ 3 : ( 0.826 , 0.265 , 0.122 , 0.323 )
G 2 ˙ 1 : ( 0.973 , 0.543 , 0.246 , 0.652 )
˙ 2 : ( 0.822 , 0.642 , 0.223 , 0.524 )
˙ 3 : ( 0.752 , 0.275 , 0.122 , 0.233 )
Table 9. Degeneration of LDFSR approximation operators into different rough set models.
Table 9. Degeneration of LDFSR approximation operators into different rough set models.
Approximation SpaceSet TheoriesFamily of SetsDegeneration of LDFSRAfter Degeneration of the
Approximation OperatorsConstructed Model
LDFS Y D = { ˙ , T ¨ ( ˙ ) : ˙ G ˙ } F ( G ˙ ) F ( G ˙ ) collection of allYesSoft fuzzy
( Q ¨ , G ˙ , ð ˜ ) fuzzy subsets of G ˙ rough sets [42,43]
LDFS Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } I ( G ˙ ) I ( G ˙ ) collection of allYesIntuitionistic fuzzy
( Q ¨ , G ˙ , ð ˜ ) IF-subsets of G ˙ soft rough sets [26]
LDFS Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } P ( G ˙ ) P ( G ˙ ) collection of allYesPythagorean fuzzy
( Q ¨ , G ˙ , ð ˜ ) PF-subsets of G ˙ soft rough sets [44]
LDFS Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) : ˙ G ˙ } R ( G ˙ ) R ( G ˙ ) collection of allYesq-rung orthopair
( Q ¨ , G ˙ , ð ˜ ) PF-subsets of G ˙ fuzzy soft rough sets
Crisp soft Y D = { ˙ , T ¨ ( ˙ ) , S ¨ ( ˙ ) , α ( ˙ ) , β ( ˙ ) : D ( G ˙ ) collection of allYesSoft rough linear
( Q ¨ , G ˙ , A ˜ ) ˙ G ˙ } D ( G ˙ ) LDF-subsets of G ˙ Diophantine fuzzy sets
Table 10. Properties of selected attributes.
Table 10. Properties of selected attributes.
AttributesCharacteristics for LDFSR
“Technical: convenience, maintainability, safety required” ( membership , non-membership , high , low )
“Monetary: operational cost, maintenance cost, purchasing cost” ( membership , non-membership , expansive , cheap )
“Operational: fuel consumption, moving speed, capacity” ( membership , non-membership , high , low )
“Strategic: flexibility, level of training required, guarantee” ( membership , non-membership , high , low )
Table 11. LDFSR.
Table 11. LDFSR.
ð ˜ ˙ 1 ˙ 2 ˙ 3 ˙ 4
G 1 ( 0.73 , 0.41 , 0.31 , 0.13 ) ( 0.63 , 0.53 , 0.13 , 0.23 ) ( 0.73 , 0.41 , 0.23 , 0.15 ) ( 0.63 , 0.53 , 0.31 , 0.36 )
G 2 ( 0.63 , 0.43 , 0.41 , 0.42 ) ( 0.74 , 0.32 , 0.63 , 0.21 ) ( 0.68 , 0.41 , 0.53 , 0.21 ) ( 0.71 , 0.41 , 0.43 , 0.28 )
G 3 ( 0.71 , 0.34 , 0.51 , 0.31 ) ( 0.63 , 0.51 , 0.43 , 0.39 ) ( 0.71 , 0.41 , 0.31 , 0.41 ) ( 0.69 , 0.38 , 0.41 , 0.31 )
G 4 ( 0.69 , 0.59 , 0.61 , 0.21 ) ( 0.81 , 0.51 , 0.31 , 0.42 ) ( 0.83 , 0.41 , 0.32 , 0.41 ) ( 0.73 , 0.49 , 0.41 , 0.21 )
G 5 ( 0.72 , 0.41 , 0.51 , 0.21 ) ( 0.83 , 0.41 , 0.42 , 0.31 ) ( 0.73 , 0.41 , 0.31 , 0.42 ) ( 0.83 , 0.49 , 0.28 , 0.41 )
G 6 ( 0.63 , 0.59 , 0.41 , 0.31 ) ( 0.78 , 0.43 , 0.38 , 0.41 ) ( 0.63 , 0.48 , 0.28 , 0.17 ) ( 0.58 , 0.49 , 0.31 , 0.42 )
G 7 ( 0.81 , 0.58 , 0.49 , 0.31 ) ( 0.73 , 0.68 , 0.43 , 0.49 ) ( 0.69 , 0.73 , 0.31 , 0.31 ) ( 0.68 , 0.51 , 0.43 , 0.21 )
Table 12. Ranking of alternatives for different score values.
Table 12. Ranking of alternatives for different score values.
LDFS G 1 G 2 G 3 G 4 G 5 G 6 G 7 RankingFinal Decision
L 1 (SF) 0.640 0.590 0.550 0.560 0.601 0.608 0.575 G 1 G 6 G 5 G 2 G 7 G 4 G 3 G 1
L 2 (QSF) 0.665 0.603 0.576 0.565 0.635 0.631 0.596 G 1 G 5 G 6 G 2 G 7 G 3 G 4 G 1
L 3 (ESF) 0.820 0.795 0.775 0.780 0.800 0.804 0.787 G 1 G 6 G 5 G 2 G 7 G 4 G 3 G 1
Table 13. The criteria for the final decision (F.D).
Table 13. The criteria for the final decision (F.D).
L ^ L * F.D
0NONO
1YESYES
0YESNO
1NONO
Table 14. Upper reduct for expert-X ( U X ) from ð ˜ ( B D ) .
Table 14. Upper reduct for expert-X ( U X ) from ð ˜ ( B D ) .
( U X ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.73 0.51 0.31 0.64 0.472 1 L 3 < 0.520 NONO
G 2 0.71 0.57 0.51 0.58 0.517 0 L 3 < 0.520 NONO
G 3 0.71 0.51 0.41 0.59 0.505 1 L 3 < 0.520 NONO
G 4 0.75 0.41 0.41 0.58 0.542 0 L 3 > 0.520 YESNO
G 5 0.83 0.51 0.41 0.58 0.537 1 L 3 > 0.520 YESYES
G 6 0.71 0.41 0.38 0.58 0.525 1 L 3 > 0.520 YESYES
G 7 0.71 0.42 0.41 0.51 0.547 1 L 3 > 0.520 YESYES
Table 15. Lower reduct for expert-X ( L X ) from ð ˜ ( B D ) .
Table 15. Lower reduct for expert-X ( L X ) from ð ˜ ( B D ) .
( L X ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.63 0.51 0.69 0.23 0.645 1 L 3 > 0.572 YESYES
G 2 0.63 0.41 0.41 0.33 0.575 0 L 3 > 0.572 YESNO
G 3 0.63 0.51 0.49 0.38 0.557 1 L 3 < 0.572 NONO
G 4 0.63 0.51 0.39 0.38 0.532 0 L 3 < 0.572 NONO
G 5 0.63 0.49 0.49 0.32 0.577 1 L 3 > 0.572 YESYES
G 6 0.63 0.49 0.59 0.38 0.587 1 L 3 > 0.572 YESYES
G 7 0.63 0.63 0.51 0.38 0.532 1 L 3 < 0.572 NONO
Table 16. Upper reduct for expert-Y ( U Y ) from ð ˜ ( B D ) .
Table 16. Upper reduct for expert-Y ( U Y ) from ð ˜ ( B D ) .
( U Y ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.73 0.51 0.31 0.64 0.472 0 L 3 < 0.520 NONO
G 2 0.71 0.57 0.51 0.58 0.517 1 L 3 < 0.520 NONO
G 3 0.71 0.51 0.41 0.59 0.505 0 L 3 < 0.520 NONO
G 4 0.75 0.41 0.41 0.58 0.542 1 L 3 > 0.520 YESYES
G 5 0.83 0.51 0.41 0.58 0.537 1 L 3 > 0.520 YESYES
G 6 0.71 0.41 0.38 0.58 0.525 1 L 3 > 0.520 YESYES
G 7 0.71 0.42 0.41 0.51 0.547 0 L 3 > 0.520 YESNO
Table 17. Lower reduct for expert-Y ( L Y ) from ð ˜ ( B D ) .
Table 17. Lower reduct for expert-Y ( L Y ) from ð ˜ ( B D ) .
( L Y ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.63 0.51 0.69 0.23 0.645 0 L 3 > 0.572 YESNO
G 2 0.63 0.41 0.41 0.33 0.575 1 L 3 > 0.572 YESYES
G 3 0.63 0.51 0.49 0.38 0.557 0 L 3 < 0.572 NONO
G 4 0.63 0.51 0.39 0.38 0.532 1 L 3 < 0.572 NONO
G 5 0.63 0.49 0.49 0.32 0.577 1 L 3 > 0.572 YESYES
G 6 0.63 0.49 0.59 0.38 0.587 1 L 3 > 0.572 YESYES
G 7 0.63 0.63 0.51 0.38 0.532 0 L 3 < 0.572 NONO
Table 18. Upper reduct for expert-Z ( U Z ) from ð ˜ ( B D ) .
Table 18. Upper reduct for expert-Z ( U Z ) from ð ˜ ( B D ) .
( U Y ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.73 0.51 0.31 0.64 0.472 1 L 3 < 0.520 NONO
G 2 0.71 0.57 0.51 0.58 0.517 1 L 3 < 0.520 NONO
G 3 0.71 0.51 0.41 0.59 0.505 0 L 3 < 0.520 NONO
G 4 0.75 0.41 0.41 0.58 0.542 0 L 3 > 0.520 YESNO
G 5 0.83 0.51 0.41 0.58 0.537 1 L 3 > 0.520 YESYES
G 6 0.71 0.41 0.38 0.58 0.525 0 L 3 > 0.520 YESNO
G 7 0.71 0.42 0.41 0.51 0.547 1 L 3 > 0.520 YESYES
Table 19. Lower reduct for expert-Z ( L Z ) from ð ˜ ( B D ) .
Table 19. Lower reduct for expert-Z ( L Z ) from ð ˜ ( B D ) .
( L Z ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.63 0.51 0.69 0.23 0.645 1 L 3 > 0.572 YESYES
G 2 0.63 0.41 0.41 0.33 0.575 1 L 3 > 0.572 YESYES
G 3 0.63 0.51 0.49 0.38 0.557 0 L 3 < 0.572 NONO
G 4 0.63 0.51 0.39 0.38 0.532 0 L 3 < 0.572 NONO
G 5 0.63 0.49 0.49 0.32 0.577 1 L 3 > 0.572 YESYES
G 6 0.63 0.49 0.59 0.38 0.587 0 L 3 > 0.572 YESNO
G 7 0.63 0.63 0.51 0.38 0.532 1 L 3 < 0.572 NONO
Table 20. Crisp soft relation A ˜ .
Table 20. Crisp soft relation A ˜ .
A ˜ ˙ 1 ˙ 2 ˙ 3 ˙ 4
G 1 0011
G 2 0110
G 3 1100
G 4 1110
G 5 1101
G 6 1001
G 7 0110
Table 21. Ranking of alternatives for different score values.
Table 21. Ranking of alternatives for different score values.
LDFS G 1 G 2 G 3 G 4 G 5 G 6 G 7 RankingFinal Decision
L 1 (SF) 0.639 0.597 0.575 0.594 0.620 0.625 0.597 G 1 G 6 G 5 G 2 = G 7 G 4 G 3 G 1
L 2 (QSF) 0.656 0.622 0.543 0.589 0.589 0.589 0.622 G 1 G 2 = G 7 G 5 = G 6 = G 4 G 3 G 1
L 3 (ESF) 0.819 0.798 0.787 0.797 0.810 0.812 0.798 G 1 G 6 G 5 G 2 = G 7 G 4 G 3 G 1
Table 22. Upper reduct for expert- X ˙ , ( U X ˙ ) from A ˜ ( H ) .
Table 22. Upper reduct for expert- X ˙ , ( U X ˙ ) from A ˜ ( H ) .
( U X ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.83 0.51 0.51 0.21 0.655 1 L 3 > 0.629 YESNO
G 2 0.75 0.51 0.51 0.32 0.607 1 L 3 < 0.629 NONO
G 3 0.71 0.41 0.41 0.33 0.595 0 L 3 < 0.629 NONO
G 4 0.75 0.41 0.51 0.32 0.632 0 L 3 > 0.629 YESNO
G 5 0.83 0.41 0.41 0.21 0.655 1 L 3 > 0.629 YESYES
G 6 0.83 0.41 0.41 0.21 0.655 1 L 3 > 0.629 YESYES
G 7 0.75 0.51 0.51 0.32 0.607 0 L 3 < 0.629 NONO
Table 23. Lower reduct for expert- X ˙ ( L X ˙ ) from A ˜ ( H ) .
Table 23. Lower reduct for expert- X ˙ ( L X ˙ ) from A ˜ ( H ) .
( L X ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.75 0.63 0.41 0.32 0.552 1 L 3 > 0.519 YESYES
G 2 0.71 0.63 0.41 0.38 0.527 1 L 3 > 0.519 YESYES
G 3 0.63 0.51 0.31 0.38 0.512 1 L 3 < 0.519 NONO
G 4 0.63 0.63 0.31 0.38 0.482 0 L 3 < 0.519 NONO
G 5 0.63 0.51 0.31 0.38 0.512 1 L 3 < 0.519 NONO
G 6 0.63 0.51 0.31 0.33 0.525 1 L 3 > 0.519 YESYES
G 7 0.71 0.63 0.41 0.38 0.527 0 L 3 > 0.519 YESNO
Table 24. Upper reduct for expert- Y ˙ , ( U Y ˙ ) from A ˜ ( H ) .
Table 24. Upper reduct for expert- Y ˙ , ( U Y ˙ ) from A ˜ ( H ) .
( U Y ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.83 0.51 0.51 0.21 0.655 1 L 3 > 0.629 YESYES
G 2 0.75 0.51 0.51 0.32 0.607 0 L 3 < 0.629 NONO
G 3 0.71 0.41 0.41 0.33 0.595 0 L 3 < 0.629 NONO
G 4 0.75 0.41 0.51 0.32 0.632 1 L 3 > 0.629 YESYES
G 5 0.83 0.41 0.41 0.21 0.655 1 L 3 > 0.629 YESYES
G 6 0.83 0.41 0.41 0.21 0.655 0 L 3 > 0.629 YESNO
G 7 0.75 0.51 0.51 0.32 0.607 1 L 3 < 0.629 NONO
Table 25. Lower reduct for expert- Y ˙ ( L Y ˙ ) from A ˜ ( H ) .
Table 25. Lower reduct for expert- Y ˙ ( L Y ˙ ) from A ˜ ( H ) .
( L Y ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.75 0.63 0.41 0.32 0.552 1 L 3 > 0.519 YESYES
G 2 0.71 0.63 0.41 0.38 0.527 0 L 3 > 0.519 YESNO
G 3 0.63 0.51 0.31 0.38 0.512 0 L 3 < 0.519 NONO
G 4 0.63 0.63 0.31 0.38 0.482 1 L 3 < 0.519 NONO
G 5 0.63 0.51 0.31 0.38 0.512 1 L 3 < 0.519 NONO
G 6 0.63 0.51 0.31 0.33 0.525 0 L 3 > 0.519 YESNO
G 7 0.71 0.63 0.41 0.38 0.527 1 L 3 > 0.519 YESYES
Table 26. Upper reduct for expert- Z ˙ , ( U Z ˙ ) from A ˜ ( H ) .
Table 26. Upper reduct for expert- Z ˙ , ( U Z ˙ ) from A ˜ ( H ) .
( U Z ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.83 0.51 0.51 0.21 0.655 1 L 3 > 0.629 YESYES
G 2 0.75 0.51 0.51 0.32 0.607 1 L 3 < 0.629 NONO
G 3 0.71 0.41 0.41 0.33 0.595 0 L 3 < 0.629 NONO
G 4 0.75 0.41 0.51 0.32 0.632 1 L 3 > 0.629 YESYES
G 5 0.83 0.41 0.41 0.21 0.655 0 L 3 > 0.629 YESNO
G 6 0.83 0.41 0.41 0.21 0.655 1 L 3 > 0.629 YESYES
G 7 0.75 0.51 0.51 0.32 0.607 1 L 3 < 0.629 NONO
Table 27. Lower reduct for expert- Z ˙ ( L Z ˙ ) from A ˜ ( H ) .
Table 27. Lower reduct for expert- Z ˙ ( L Z ˙ ) from A ˜ ( H ) .
( L Z ˙ ) T ¨ D S ¨ D α D β D L 3 L ^ L * F.D
G 1 0.75 0.63 0.41 0.32 0.552 1 L 3 > 0.519 YESYES
G 2 0.71 0.63 0.41 0.38 0.527 1 L 3 > 0.519 YESYES
G 3 0.63 0.51 0.31 0.38 0.512 0 L 3 < 0.519 NONO
G 4 0.63 0.63 0.31 0.38 0.482 1 L 3 < 0.519 NONO
G 5 0.63 0.51 0.31 0.38 0.512 0 L 3 < 0.519 NONO
G 6 0.63 0.51 0.31 0.33 0.525 1 L 3 > 0.519 YESYES
G 7 0.71 0.63 0.41 0.38 0.527 1 L 3 > 0.519 YESYES
Table 28. Comparison of LDFSRS and SRLDFS with the existing concepts.
Table 28. Comparison of LDFSRS and SRLDFS with the existing concepts.
ConceptsSatisfactionDissatisfactionReference
DegreeDegreeParameterizations
Fuzzy set [46]××
Rough set [55]×××
Soft set [56]×××
Intuitionistic fuzzy set [47,48]×
Pythagorean fuzzy set [49,50,51]×
q-rung orthopair fuzzy set [52,53]×
LDFS [54]
LDFSS (proposed)
LDFSRS (proposed)
SRLDFS (proposed)
ConceptsUpper and lowerBoundarymulti-valued
approximationsregionparameterizations
Fuzzy set [46]×××
Rough set [55]×
Soft set [56]××
Intuitionistic fuzzy set [47,48]×××
Pythagorean fuzzy set [49,50,51]×××
q-rung orthopair fuzzy set [52,53]×××
LDFS [54]×××
LDFSS (proposed)××
LDFSRS (proposed)
SRLDFS (proposed)
Table 29. Comparison of LDFSRS and SRLDFS with the existing concepts.
Table 29. Comparison of LDFSRS and SRLDFS with the existing concepts.
ConceptsRemarks
Fuzzy set [46]It only deals with the truth values of objects.
Rough set [55]It only deal with the vagueness of input data.
Soft set [56]It only deal with the uncertainties under parameterizations.
Intuitionistic fuzzy set [47]It cannot be applied if 1 < T ¨ I ( G ) + S ¨ I ( G ) 2 for some G .
Pythagorean fuzzy set [49,50,51]It cannot be applied if 1 < T ¨ I 2 ( G ) + S ¨ I 2 ( G ) 2 for some G .
q-rung orthopair fuzzy set [52,53]It cannot be applied for smaller values of “q” with
1 < T ¨ O q ( G ) + S ¨ O q ( G ) 2 or if T ¨ O ( G ) = S ¨ O ( G ) = 1 for some G .
LDFS [54](1) It can deal with all the cases in which FS, IFS, PFS, and q-ROFS
cannot be applied; (2) it involves a parameterization perspective
and works under the influence of reference or control parameters; (3) satisfaction
and dissatisfaction degrees can be chosen freely from [ 0 , 1 ] .
LDFSS (proposed)It contains all the properties of LDFS with the addition of multi-valued
parameterizations to deal with the uncertainties in a parametric manner.
LDFSRS (proposed)It contains all the properties of LDFSS with the addition of upper
and lower approximations to deal with the roughness of input data
under the effect of “LDFS approximation space”.
SRLDFS (proposed)It contains all the properties of LDFSS with the addition of upper
and lower approximations to deal with the roughness of input data
under the effect of “crisp soft approximation space”.
Table 30. Comparison of the results obtained from the proposed algorithms.
Table 30. Comparison of the results obtained from the proposed algorithms.
ProposedScoreCoreFinal
AlgorithmFunctionSetDecision
Algorithm 1 L 1 × G 1
Algorithm 1 L 2 × G 1
Algorithm 1 L 3 × G 1
Algorithm 2× G 5
Algorithm 3 L 1 × G 1
Algorithm 3 L 2 × G 1
Algorithm 3 L 3 × G 1
Algorithm 4× G 1

Share and Cite

MDPI and ACS Style

Riaz, M.; Hashmi, M.R.; Kalsoom, H.; Pamucar, D.; Chu, Y.-M. Linear Diophantine Fuzzy Soft Rough Sets for the Selection of Sustainable Material Handling Equipment. Symmetry 2020, 12, 1215. https://doi.org/10.3390/sym12081215

AMA Style

Riaz M, Hashmi MR, Kalsoom H, Pamucar D, Chu Y-M. Linear Diophantine Fuzzy Soft Rough Sets for the Selection of Sustainable Material Handling Equipment. Symmetry. 2020; 12(8):1215. https://doi.org/10.3390/sym12081215

Chicago/Turabian Style

Riaz, Muhammad, Masooma Raza Hashmi, Humaira Kalsoom, Dragan Pamucar, and Yu-Ming Chu. 2020. "Linear Diophantine Fuzzy Soft Rough Sets for the Selection of Sustainable Material Handling Equipment" Symmetry 12, no. 8: 1215. https://doi.org/10.3390/sym12081215

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop