You are viewing a javascript disabled version of the site. Please enable Javascript for this site to function properly.
Go to headerGo to navigationGo to searchGo to contentsGo to footer
In content section. Select this link to jump to navigation

An Entropy-Based Method for Probabilistic Linguistic Group Decision Making and its Application of Selecting Car Sharing Platforms

Abstract

As the tourism and mobile internet develop, car sharing is becoming more and more popular. How to select an appropriate car sharing platform is an important issue to tourists. The car sharing platform selection can be regarded as a kind of multi-attribute group decision making (MAGDM) problems. The probabilistic linguistic term set (PLTS) is a powerful tool to express tourists’ evaluations in the car sharing platform selection. This paper develops a probabilistic linguistic group decision making method for selecting a suitable car sharing platform. First, two aggregation operators of PLTSs are proposed. Subsequently, a fuzzy entropy and a hesitancy entropy of a PLTS are developed to measure the fuzziness and hesitancy of a PLTS, respectively. Combining the fuzzy entropy and hesitancy entropy, a total entropy of a PLTS is generated. Furthermore, a cross entropy between PLTSs is proposed as well. Using the total entropy and cross entropy, DMs’ weights and attribute weights are determined, respectively. By defining preference functions with PLTSs, an improved PL-PROMETHEE approach is developed to rank alternatives. Thereby, a novel method is proposed for solving MAGDM with PLTSs. A car sharing platform selection is examined at length to show the application and superiority of the proposed method.

1Introduction

With the development of mobile internet and “Internet +”, sharing economy, such as traffic-sharing, health care-sharing and food-sharing, has boomed in recent years. Car-sharing, a new mode of traffic-sharing, is becoming more and more popular along with the development of tourism. People usually complete the car rental and provide their feedbacks on consumer experiences on sharing platforms. Therefore, how to select a suitable car sharing platform is important for tourists. Since the car sharing platforms are often evaluated in safety, convenience, the brand of car and so on, the selection of car sharing platform can be described as a kind of multi-attribute decision making problems. Ordinarily, several tourists travel together. Hence, more than one tourist (decision maker, shorted by DM) decides which car sharing platform is selected. Thereby, the car sharing platform selection can be considered as a kind of multi-attribute group decision making (MAGDM) (Xu et al.2016; Wan et al.2015; Dong et al.2018; Kou et al.2020; Zhang et al.2019).

Owing to the complexity of problems and vagueness of human thinking, it is difficult for DMs to describe the decision information as crisp numbers. DMs are more apt to use linguistic variables (Zadeh, 1975) to evaluate alternatives. However, the linguistic variable allows DMs to express decision information with only one linguistic term (LT). Sometimes, DMs cannot adequately describe their evaluations using one exact LT and have hesitancy among several LTs. To overcome this drawback of the linguistic variable, Rodríguez et al. (2012) introduced the hesitant fuzzy linguistic term set (HFLTS) which permits DMs to express their preferences on alternatives with several possible LTs. Nevertheless, all possible LTs in a HFLTS have the same importance. In fact, DMs may prefer a LT to others. Thus, some useful information may be lost (Pang et al.2016). To make up for this limitation, Pang et al. (2016) generalized HFLTS and presented the probabilistic linguistic term set (PLTS) in which each possible LT is assigned a probability (weight). Hence, the PLTS contains more useful information compared with the HFLTS. In recent years, the PLTS has gained more attention and some research achievements have been obtained. These achievements can be roughly divided into four categories.

(1) Operations and aggregation operators of PLTSs. Operations of PLTSs and aggregation operators are important to aggregate decision information in the PLTS circumstance. Pang et al. (2016) firstly defined some operational laws of PLTSs, and then put forward PLWA and PLWG operators of PLTSs, respectively. Subsequently, Zhang et al. (2017) pointed that the operation result by the laws (Pang et al.2016) is a linguistic term rather than a PLTS. To remedy this defect, Zhang et al. (2017) defined new operations of PLTSs. Nevertheless, the operation results by these new operations in Zhang et al. (2017) may exceed the bounds of LTSs. To overcome this limitation, Gou and Xu (2016) introduced novel operations of PLTSs by a linguistic scale function. Later, using this function, Mao et al. (2019) introduced new operation laws and developed the GPLHWA operator and GPLHOWA operator based on Archimedean t-norms and s-norms. Zhang (2018) investigated the large group decision making and proposed a PL-WAA operator. Recently, diverse types of operators, such as power operators (Liu et al., 2019) and dependent weighted average operators (Liu et al., 2019), have been proposed one after another. Furthermore, Mi et al. (2020) conducted a survey of existing achievements on operations of PLTSs and aggregation operators.

(2) Distance measure of PLTSs. Pang et al. (2016) and Zhang et al. (2016) defined Euclidean distance and Hamming distance measures, respectively, of PLTSs based on the probabilities and subscripts of possible LTs. Afterwards, Wu and Liao (2018) argued that the results derived by these two distance measures (Pang et al.2016; Zhang et al.2016) occasionally are against human intuition. To overcome this defect, they proposed an improved distance measure. However, the calculation of this improved distance measure is very complex. Lately, Mao et al. (2019) proposed a simple Euclidean distance measure by a linguistic scale function (Gou and Xu, 2016). Unfortunately, some counterintuitive results still appear based on this Euclidean distance, please see Table 2 in Section 3.2.

(3) Entropy and cross entropy of PLTSs. Due to the uncertainty of information with PLTSs, how to measure such a certainty is important. Entropy of PLTSs, as an efficient tool to measure the uncertainty of information, is an interesting topic but has not gained wide attention. Only Lin et al. (2019) developed an information entropy of PLTSs by probabilities of possible LTs in a PLTS. Liu et al. (2018) proposed three kinds of entropy of PLTSs by extending the entropy of HFSs into the PLTS context, including the fuzzy entropy, hesitancy entropy and total entropy. On the other hand, the cross entropy is effective for measuring the differences between PLTSs. Up to now, only Liu and Teng (2019) presented two cross entropy measures of PLTSs based on sine and tangent trigonometric functions.

(4) Decision methods for solving MAGDM problems with PLTSs. It is important to choose proper decision methods for selecting best alternatives. At present, pools of methods have been proposed to solve MAGDM problems with PLTSs. For example, Pang et al. (2016) presented a distance-based extended TOPSIS method with PLTSs. Zhang X.F. et al. (2019) argued that this extended TOPSIS only considered the distance proximity of alternatives with respect to ideal and negative ideal solutions, while ignored the proximity in direction. In response to this problem, Zhang X.F. et al. (2019) developed a projection method to solve MAGDM with PLTSs. Later, by improving the classical MULTIMOORA method in different angles, Wu et al. (2018) and Liu and Li (2019) developed a PL-MULTIMOORA method and an extended MULTIMOORA method, respectively. Unfortunately, these two methods did not take negative ideal solutions into account. Considering both positive and negative ideal solutions, Li et al. (2020) put forward a PP-MULTIMOORA method. In addition, diverse types of outranking methods (Lin et al.2019; Liao et al.2019; Xu et al.2019; Peng et al.2020; Liu and Li, 2018), such as ELECTRE and PROMETHEE, are also presented. To facilitate the study on decision methods of PLTSs, Liao et al. (2020) provided a survey which includes most decision methods and their applications.

Although many achievements have been achieved, they suffer from some limitations.

(1) Though the GPLHWA operator (Mao et al.2019) improved the PLWA operator (Pang et al.2016) and the PL-WAA operator (Zhang et al.2017) to some extent, it has some flaws that the aggregated result obtained by the GPLHWA operator is not a PLTS in a strict sense. The reason is that the sum of probabilities of all possible LTs in the aggregated result is more than 1. Therefore, proposing new aggregation operators with desirable properties is helpful for aggregating evaluation values of alternatives.

(2) By existing distance measures of PLTSs (Pang et al.2016; Mao et al.2019; Zhang et al.2016), some counter-intuitive results appears. Although the results derived by the distance measure (Wu and Liao, 2018) almost agree with human intuition, the computation of this distance measure is too complex to use easily. Therefore, it is necessary to develop a new simple distance measure of PLTSs by which computation results coincide with human intuition.

(3) The research on entropy and cross entropy of PLTSs is very little. Up to date, only Lin et al. (2019) and Liu et al. (2018) addressed the entropy of PLTSs. However, the distinguishing power of the entropy proposed by Lin et al. (2019) is not high enough. Although the fuzzy entropy and hesitancy entropy proposed by Liu et al. (2018) can neatly measure the uncertainty of PLTSs, some properties of this fuzzy entropy are counter-intuitive and the computation of this hesitancy entropy is not simple. In addition, the cross entropy proposed by Liu and Teng (2019) fails for the symmetric linguistic term sets in PLTSs. Thereby, it is valuable to study deeply on the entropy and cross entropy of PLTSs.

(4) Existing decision methods are fruitful to solve decision problems with PLTSs. Nevertheless, some methods (Lin et al.2019; Xu et al.2019; Li et al.2020) can only solve MADM problems, but fail for MAGDM problems. Although these methods (Pang et al.2016; Wu et al.2018; Zhang X.F. et al.2019; Liu et al., 2019) are capable to solve MAGDM problems with PLTSs, DMs’ weights or attribute weights are not considered or assigned in advance, which may result in subjective random weights. Hence, it is interesting to seek a novel method which not only can solve MAGDM problems, but also determine DMs’ weights and attribute weights objectively.

To make up for above limitations, this paper proposes a novel method for solving MAGDM problems with PLTSs. First, two aggregation operators, including PLWAM (probabilistic linguistic weighted arithmetic mean) operator and PLWGM (probabilistic linguistic weighted geometric mean) operator, respectively, are proposed and some desirable properties are studied. To measure the hesitancy degree of a PLTS, a hesitancy index of the PLTS is introduced. Then a general distance measure of PLTSs is defined to measure the deviation between two PLTSs. Considering the fact that the uncertainty of a PLTS includes the fuzziness and hesitancy, a fuzzy entropy and hesitancy entropy of a PLTS are defined and then a total entropy of a PLTS is derived to measure such uncertainties. Meanwhile, a cross entropy of PLTSs is defined to measure the distinction between PLTSs. Afterwards, by minimizing the total entropy and the cross entropy of a DM, an objective program model is built to determine DMs’ weights. Subsequently, individual decision matrices are aggregated into a collective one by the PLWAM operator. To derive attributes weights objectively, a bi-objective program is built by minimizing the total entropy of attribute values with respect to each attribute as well as maximizing the cross entropy between attribute values of alternatives. By defining a new preference function in the form of PLTSs, an improved PL-PROMETHEE method is developed to rank alternatives. Thereby, a novel method is proposed for solving MAGDM problems with PLTSs. A case of car sharing platform selection is applied to show the effectiveness and advantages of the proposed method at length. The primary features of the proposed method are outlined as follows:

(1) Two new probabilistic linguistic average aggregation operators of PLTSs (i.e. PLWAM and PLWAG operators) are proposed. A prominent characteristic of them is that the aggregated result obtained by these two operators is not only a PLTS with the sum of probabilities of possible LTs being equal to 1, but also is consistent with human intuition.

(2) A new generalized distance measure of PLTSs is defined. It is worth mentioning that the hesitancy degree of a PLTS is considered in this distance measure. Thus, the new distance has a stronger distinguishing power. Moreover, a ranking approach is presented to rank PLTSs.

(3) A fuzzy entropy, a hesitancy entropy and a cross entropy of PLTSs are introduced. The fuzzy entropy has desirable properties and the computation of the hesitancy entropy is simple. Meanwhile, the cross entropy can distinguish the deviations between PLTSs with symmetric linguistic term sets.

(4) Based on entropy and cross entropy of PLTSs, distinct objective programs are established to determine DMs’ weights and attribute weights objectively. Finally, an improved PL-PROMETHEE method is developed to rank alternatives.

The remainder of this paper is organized as follows: In Section 2, some basic concepts of PLTSs are reviewed. Moreover, PLWAM and PLWGM operators are proposed. Section 3 introduces a hesitancy index of a PLTS and then develops a generalized distance measure of PLTSs. Based on this distance measure, a ranking approach is presented to sort PLTSs. Section 4 defines several types of entropy of PLTSs, including the fuzzy entropy, hesitancy entropy, total entropy and cross entropy of PLTSs. In Section 5, an improved PL-PROMETHEE method is developed to solve MAGDM problems with PLTSs. Section 6 provides a case study of car sharing platform selection to illustrate the application of the proposed method. Furthermore, comparison analyses are conducted to show advantages of the proposed method. Some conclusions are made in Section 7.

2Preliminaries

In this section, some definitions and notions related to the PLTS are reviewed. Furthermore, two aggregation operators of PLTSs are proposed and some desirable properties of them are investigated.

Definition 1

Definition 1(See Xu, 2005).

Let S={sα|α=τ,,1,0,1,,τ} be a finite and totally ordered discrete LTS, where sα represents a possible value for a linguistic term, and τ is a positive integer. Especially, the mid-linguistic label s0 represents an assessment of “indifference”, and the rest of them are placed symmetrically around it. sτ and sτ are lower and upper bounds of linguistic labels.

To preserve all given linguistic information, Xu (2004) extended the discrete LTS S into a continuous LTS S¯={sα|α[τ,τ]. If sαS, then sα is called an original linguistic term; Otherwise, sα is called a virtual linguistic term.

Definition 2

Definition 2(See Gou and Xu, 2016).

Let S={sα|α=τ,,1,0,1,,τ} be a LTS, the linguistic term sα that expresses the equivalent information to the membership degree γ is obtained by a linguistic scale function g:

g:[sτ,sτ][0,1],g(sα)=α+τ2τ=γ.

Additionally, the membership degree γ that expresses the equivalent information to the linguistic term sα is obtained by the following function

g1:[0,1][sτ,sτ],g1(γ)=s(2γ1)τ=sα.

Definition 3

Definition 3(See Pang et al.2016).

Let S={sα|α=τ,,1,0,1,,τ} be a linguistic term set, a PLTS is defined as

L(p)={L(k)(p(k))|L(k)S,p(k)0,k=1,2,,#L(p),k=1#L(p)p(k)1},
where L(k)(p(k)) represents the linguistic term L(k) associated with the probability p(k), and #L(p) is the number of all different linguistic terms in L(p).

For a PLTS L(p) with k=1#L(p)p(k)1, Pang et al. (2016) gave a normalizing method.

Definition 4

Definition 4(See Pang et al.2016).

Given a PLTS L(p) with k=1#L(p)p(k)1, the normalized PLTS L˙(p) is defined as:

L˙(p)={L(k)(p˙(k))|k=1,2,,#L(p)},
where p˙(k)=p(k)/k=1#L(p)p(k) for all k=1,2,,#L(p).

To ensure that the operational results among PLTSs can be straightforwardly determined, Mao et al. (2019) defined an ascending ordered PLTS below.

Definition 5.

Given a PLTS L(p)={L(k)(p(k))|L(k)S,p(k)0,k=1,2,,#L(p)}, where r(k) is the subscript of linguistic term L(k), an ascending ordered PLTS can be derived by the following steps:

  • (1) If all elements in a PLTS are with different values of r(k)p(k), then all elements are arranged according to the value of r(k)p(k) (k=1,2,,#L(p)) in an ascending order;

  • (2) If two or more elements with equal values of r(k)p(k), then

    • (a) When the subscripts r(k) (k=1,2,,#L(p)) are unequal, r(k)p(k) (k=1,2,,#L(p)) are arranged according to values of r(k) (k=1,2,,#L(p)) in an ascending order;

    • (b) When the subscripts r(k) (k=1,2,,#L(p)) are equal, r(k)p(k) (k=1,2,,#L(p)) are arranged according to values of p(k) (k=1,2,,#L(p)) in an ascending order.

If a PLTS L(p) is normalized by Definition 4 and all elements of L(p) is ordered by Definition 5, then L(p) is converted into a normalized ordered PLTS L¯(p)={L¯(k)(p¯(k))|k=1,2,,#L¯(p)}.

In real decision making problems, the numbers of linguistic terms in two different PLTSs are often different. This makes trouble to operate. In order to make them have the same number of linguistic terms, Pang et al. (2016) provided a method to add the number of linguistic terms for PLTSs in which the number of linguistic terms is relatively small as follows.

Definition 6.

Let L1(p) and L2(p) be two PLTSs, where L1(p)={L1(k)(p1(k))|k=1,2,,#L1(p)} and L2(p)={L2(k)(p2(k))|k=1,2,,#L2(p)}, #L1(p) and #L2(p) are the numbers of linguistic terms in L1(p) and L2(p), respectively. If #L1(p)>#L2(p), then add #L1(p)#L2(p) linguistic terms to L2(p). The added linguistic terms are the smallest linguistic terms in L2(p) and the probabilities of added linguistic terms are zero.

Definition 7

Definition 7(See Mao et al.2019).

Let L¯1(p) and L¯2(p) be two PLTSs, where L¯1(p)={L¯1(k)(p¯1(k))|k=1,2,,#L¯1(p)}, L¯2(p)={L¯2(k)(p¯2(k))|k=1,2,,#L¯2(p)} and #L¯1(p)=#L¯2(p). Then the Euclidean distance d(L¯1(p),L¯2(p)) between L¯1(p) and L¯2(p) is defined as

(1)
d(L¯1(p),L¯2(p))=k=1#L¯1(p)(p¯1(k)g(L¯1(k))p¯2(k)g(L¯2(k)))/#L¯1(p).

Definition 8

Definition 8(See Wu and Liao, 2018).

Given a PLTS L(p)={L(k)(p(k))|L(k)S,p(k)0,k=1,2,,#L(p)}, the expected function of L(p) is defined as

(2)
e(L(p))=k=1#L(p)g(L(k))p(k)/k=1#L(p)p(k).

2.1Aggregation Operators of PLTSs

This subsection reviews the operational laws of PLTSs presented by Gou and Xu (2016), and then proposes PLWAM and PLWGM aggregation operators, which are used to aggregate the decision information in GDM discussed in later sections.

Definition 9

Definition 9(See Gou and Xu, 2016).

Let S={sα|α=τ,,1,0,1,,τ}, L(p), L1(p) and L2(p) be three PLTSs, and λ be a positive real number, it stipulates:

  • (i) L1(p)L2(p)=g1(k1=1,2,,#L1(p)k1=1,2,,#L2(p){(g(L1(k1))+g(L2(k2))g(L1(k1))g(L2(k2)))(p1(k1)p2(k2))});

  • (ii) L1(p)L2(p)=g1(k1=1,2,,#L1(p)k1=1,2,,#L2(p){(g(L1(k1))g(L2(k2)))(p1(k1)p2(k2))});

  • (iii) λL(p)=g1(k=1,2,,#L(p){1(1g(L(k)))λ,(p(k))});

  • (iv) (L(p))λ=g1(k=1,2,,#L(p){(g(L(k)))λ,(p(k))});

  • (v) The complement of L(p): (L(p))(c)=g1(k=1,2,,#L(p){(1g(L(k)))(p(k))}).

Remark 1.

As reviewed in Mi et al. (2020), up to now existing operations of PLTSs can be divided into five types: symbolic-based computation (Pang et al.2016), semantic-based computation (Gou and Xu, 2016; Wu and Liao, 2019), triangular norms-based computation (Klement et al.2000), evidence reasoning-based computation (Li and Wei, 2019) and double alpha-cut-based computation (Jiang and Liao, 2020). Among these operations, only the semantic-based computation (Gou and Xu, 2016; Wu and Liao, 2019) and the evidence reasoning-based computation (Li and Wei, 2019) satisfy the closure of operation of PLTSs. The semantic-based computations (Gou and Xu, 2016; Wu and Liao, 2019) defined common operations, such as additive, subtraction and numerical operations of PLTSs. However, the computation (Li and Wei, 2019) only defined the additive operation of PLTSs. To sum up, the semantic-based computation (Gou and Xu, 2016; Wu and Liao, 2019) is more reasonable compared with other computations. Therefore, it is natural to choose this kind of computation to complete the operations of PLTSs. The differences between operations in Gou and Xu (2016) and Wu and Liao (2019) are the selections of linguistic scale functions and the tools for handling two PLTSs with different numbers of LTs. Compared with the computation in Wu and Liao (2019) (see Definition 3 in Wu and Liao, 2019), the computation in Gou and Xu (2016) is easier to operate and more widely used. For example, many achievements (Bai et al.2017; Liang et al.2018; Feng et al.2020) have been gained based on the computation in Gou and Xu (2016), while only one method (Wu et al.2019) was developed by the computation in Wu and Liao (2019). At this point, we choose the computation in Gou and Xu (2016) to conduct the operations of PLTSs in this paper.

In virtue of Definition 9, a PLWAM operator of PLTS is presented.

Definition 10.

Let Li(p)={Li(k)(pi(k))|k=1,2,,#Li(p)} (i=1,2,,n) be n PLTSs. A PLWAM operator is a function fnf such that:

(3)
PLWAM(L1(p),L2(p),,Ln(p))=ω1L1(p)ω2L2(p)ωnLn(p),
where ω=(ω1,ω2,,ωn)T is the weight vector of Li(p) (i=1,2,,n), satisfying 0ωi1(i=1,2,,n) and i=1nωi=1.

Theorem 1.

Given n PLTSs L1(p),L2(p),,Ln(p), the aggregated value by the PLWAM operator is also a PLTS as

(4)
PLWAM(L1(p),L2(p),,Ln(p))=g1(kj=1,2,,#Lj(p)j=1,2,,n{(1j=1n(1g(Lj(kj)))ωj)(p1(k1)p2(k2)pn(kn))}).
Furthermore,
(5)
kj=1,2,,#Lj(p)j=1,2,,np1(k1)p2(k2)pn(kn)=1.

Proof.

Please see Appendix A.  □

Analogously, a PLWGM operator of PLTSs is presented.

Definition 11.

Let Li(p)={Lik(p)|k=1,2,,#Li(p)} (i=1,2,,n) be n PTTSs. A PLWGM operator of PLTSs is a function fnf such that:

(6)
PLWGM(L1(p),L2(p),,Ln(p))=(L1(p))ω1(L2(p))ω2(Ln(p))ωn,
where ω=(ω1,ω2,,ωn)T is the weight vector of Li(p) (i=1,2,,n), satisfying 0ωi1(i=1,2,,n) and i=1nωi=1.

In a way similar to the proving process of Theorem 1, Theorem 2 can be proved based on the operational laws (ii) and (iv) in Definition 9.

Theorem 2.

Given n PLTSs L1(p),L2(p),,Ln(p), the aggregated value by the PLWGM operator is also a PLTS as follows

(7)
PLWGM(L1(p),L2(p),,Ln(p))=g1(kj=1,2,,#Lj(p)j=1,2,,n{(j=1n(g(Lj(kj)))ωj)(p1(k1)p2(k2)pn(kn))}),
and
(8)
kj=1,2,,#Lj(p)j=1,2,,np1(k1)p2(k2)pn(kn)=1.

Remark 2.

Compared with existing aggregation operators (Pang et al.2016; Mao et al.2019; Zhang, 2018), the proposed operators have significant advantages. For instance, they satisfy the closure of operations and the computation results are closer to human intuition. These advantages can be verified by Example 1 in what follows.

Example 1.

Given a LTS S={s3,s2,s1,s0,s1,s2,s3}. Let L1(p)={s2(0.4),s3(0.4)}, L2(p)={s1(0.1),s2(0.6),s3(0.1)} and L3(p)={s2(0.2),s3(0.1)} be three PLTSs. The weight vector is w=(0.3,0.2,0.5)T. By applying existing arithmetic weighted aggregation operators and the proposed PLWAM operator, aggregated results are obtained and shown in Table 1.

Table 1

Results obtained by different aggregation operators.

OperatorsResultsPLTSInformation lostInformation distortion
PLWA (Pang et al.2016){s0,s0.2,s0.4,s0.6,s0.75,s0.95,s1.15,s1.5,s1.7,s2.25}NoYesYes
GPLHWA (Mao et al.2019){s1.85(0.14),s2.00(0.16),s3(1.36)}NoYesYes
PL-WAA
(Zhang, 2018){s3(0.064),s2(0.064),s1(0.064),s0(0.064),s1(0.084),s2(0.424),s3(0.254)}YesYesYes
PLWAM in this paper{s1.85(0.04),s2.00(0.25),s3(0.71)}YesNoNo

It is observed from Table 1 that the PLWAM operator proposed in this paper has following merits:

(1) The PLWAM operator satisfies the closure of operations, i.e. the aggregated result of PLTSs obtained by the PLWAM operator is still a PLTS. However, the aggregated result obtained by the PLWA operator (Pang et al.2016) is a linguistic term rather than a PLTS. Thus, the probability information is lost. Although the aggregated result obtained by the GPLHWA operator (Mao et al.2019) seems to be a PLTS, the sum of probabilities of all possible LTs in this result is more than 1. Therefore, this result is not a PLTS in a strict sense. Hence, there exists information distortion while using the GPLHWA operator.

(2) The aggregated result obtained by the proposed PLWAM operator is more consistent with human intuition. Observing L1(p), L2(p) and L3(p), it is deduced that the aggregated result should be larger than s1 and smaller than s3. Furthermore, the probability of s3 should be more than that of s1 because three PLTSs include s3 but only L2(p) contains s1 with a small probability. By the PLWAM operator proposed in this paper, the aggregated result (see Table 1) is in accordance with this analysis indeed. On the other hand, the aggregated result obtained by the PL-WAA operator (Zhang, 2018) is counterintuitive because LTs s3, s2 and s1 are all included in this result, which means that the information is distorted in aggregating process.

(3) Observing Table 1, the aggregated results obtained by the proposed PLWAM operator and the PL-WAA (Zhang, 2018) operator are both PLTSs. Nevertheless, this paper proves that all the aggregated results by the PLWAM operator are all PLTSs for any PLTSs (see Theorem 1), but Zhang (2018) failed to do so.

Analogously, the PLWGM operator has above advantages.

3A Hesitancy Index of a PLTS and an Approach to Ranking PLTSs

In this section, a hesitancy index of a PLTS is introduced to measure the hesitancy degree of a PLTS, and then a new generalized distance measure between two PLTSs is developed. Finally, based on the proposed distance measure, a TOPSIS-based approach is presented to rank PLTSs.

3.1A Hesitancy Index of a PLTS

Definition 12.

Suppose a normalized PLTS L˙(p)={L(k)(p˙(k))|k=1,2,,#L(p)}. A hesitancy index of L˙(p) is defined as

(9)
h(L˙(p))=2k=1#L˙(p)|g(L(k))g¯|p˙(k)#L˙(p)>1,0#L˙(p)=1,
where g¯=1#L˙(p)k=1#L˙(p)g(L(k)).

It is clear from Eq. (9) that the hesitancy index h(L˙(p)) satisfies 0h(L˙(p))1. Especially, h(L˙(p))=0 if L˙(p)={sα(1)|sαS} and h(L˙(p))=1 if L˙(p)={sτ(0.5),sτ(0.5)}. This result is in accordance with human intuition.

3.2A General Distance Measure Between PLTSs

First, this subsection reviews existing distance measures between PLTSs (Pang et al.2016; Mao et al.2019; Zhang et al.2016), and then develops a general distance measure between PLTSs based on the proposed hesitancy index. In the end, comparative analyses of the proposed distance measure and existing ones are conducted.

Let S={sα|α=τ,,1,0,1,,τ} be a LTS. L¯1(p) and L¯2(p) are two PLTSs, where L¯1(p)={L¯1(k)(p¯1(k))|k=1,2,,#L¯1(p)}, L¯2(p)={L¯2(k)(p¯2(k))|k=1,2,,#L¯2(p)} and #L¯1(p)=#L¯2(p)=#L¯(p).

Pang et al. (2016) defined a deviation degree between two PLTSs as

(10)
dP(L¯1(p),L¯2(p))=k=1#L¯(p)(p1(k)r1(k)p2(k)r2(k))/#L¯(p),
where r1(k) and r2(k) are the subscripts of linguistic terms of L¯1(p) and L¯2(p), respectively.

Later, Zhang et al. (2016) introduced a distance measure between PLTSs as below.

(11)
dZ(L¯1(p),L¯2(p))=k=1#L¯(p)p(r1(k),r2(k))d(r1(k),r2(k)),
where p(r1(k),r2(k))=p(r1(k))p(r2(k))=p1(k)p2(k) and d(r1(k),r2(k))=(r1(k)r2(k))/T, T is the number of linguistic terms in S.

Mao et al. (2019) argued some limitations of above distance measures and developed an Euclidean distance as

(12)
dM(L¯1(p),L¯2(p))=k=1#L¯(p)(p1(k)g(L¯1(k))p2(k)g(L¯2(k)))/#L¯(p).

This paper defines a general distance measure between PLTSs by the proposed hesitancy index.

Definition 13.

Given two PLTSs L¯1(p)={L¯1(k)(p¯1(k))|k=1,2,,#L¯1(p)} and L¯2(p)={L¯2(k)(p¯2(k))|k=1,2,,#L¯2(p)}, where #L¯1(p)=#L¯2(p)=#L¯(p), a general distance measure of PLTSs is defined as

(13)
dG(L¯1(p),L¯2(p))=(12(1#L¯(p)k=1#L(p)12(|g(L¯1(k))g(L¯2(k))|q+|p¯1(k)p¯2(k)|)q+|h(L¯1(p))h(L¯2(p)|q))1/q,
where q1.

Specially, when q=1, a Manhattan distance is obtained as

(14)
dM(L¯1(p),L¯2(p))=12(1#L¯(p)k=1#L(p)12(|g(L¯1(k))g(L¯2(k))|+|p¯1(k)p¯2(k)|)+|h(L¯1(p))h(L¯2(p)|),
when q=2, a novel Euclidean distance of PLTSs is developed as
(15)
dE(L¯1(p),L¯2(p))=12(1#L¯(p)k=1#L(p)12((g(L¯1(k))g(L¯2(k)))2+(p¯1(k)p¯2(k))2)+(h(L¯1(p))h(L¯2(p))2).

According to Eqs. (14) and (15), it is easily proved that dM and dE have desirable properties described in Theorem 3.

Theorem 3.

For two PLTSs L¯1(p) and L¯2(p) with #L¯1(p)=#L¯2(p), dM and dE are the novel Manhattan and Euclidean distances of PLTSs, respectively. They satisfy:

  • (i) 0dM(L¯1(p),L¯2(p)),dE(L¯1(p),L¯2(p))1;

  • (ii) dM(L¯1(p),L¯2(p))=dE(L¯1(p),L¯2(p))=0 if and only if L¯1(p)=L¯2(p);

  • (iii) dM(L¯1(p),L¯2(p))=dM(L¯2(p),L¯1(p)) and dE(L¯1(p),L¯2(p))=dE(L¯2(p),L¯1(p)).

Remark 3.

Compared with existing distance measures in Pang et al. (2016), Mao et al. (2019) and Zhang et al. (2016), the distance measures defined in Definition 13 have a stronger distinguishing power. Table 2 shows this advantage explicitly.

It is shown from Table 2 that PLTSs in each pair are not identical completely. Therefore, the distance between each pair PLTSs should be more than zero intuitively. The proposed distance measures do this exactly. However, the distances by existing distance measures (Pang et al.2016; Mao et al.2019; Zhang et al.2016) are all zeros, so the difference of PLTSs in each pair cannot be distinguished. This means that the proposed distance measures have stronger distinguishing powers, which can be attributed to the following factors:

Table 2

Distancesbetween PLTSs obtained by different distance measures.

PLTSsDifferent distance measuresDistances between PLTSs
L¯1(p)={s2(0.8),s3(0.2)}

L¯2(p)={s2(0.5),s3(0.5)}
Zhang’s distance (Zhang et al.2016)dZ(L¯1(p),L¯2(p))=0
The Manhattan distance proposed in this paperdM(L¯1(p),L¯2(p))=0.075>0
The Euclidean distance proposed in this paperdE(L¯1(p),L¯2(p))=0.15>0
L¯3(p)={s0(0.4),s1(0.6)} L¯4(p)={s0(0.8),s3(0.2)}Pang’s distance (Pang et al.2016)dP(L¯3(p),L¯4(p))=0
The Manhattan distance proposed in this paperdM(L¯3(p),L¯4(p))=0.3083>0
The Euclidean distance proposed in this paperdE(L¯3(p),L¯4(p))=0.3308>0
L¯5(p)={s3(0.8),s1(0.2)}

L¯6(p)={s3(0.7333),s0(0.2667)}
The Mao’s Euclidean distance (Mao et al.2019)d(L¯5(p),L¯6(p))=0
The Manhattan distance proposed in this paperdM(L¯5(p),L¯6(p))=0.1208>0
The Euclidean distance proposed in this paperdE(L¯5(p),L¯6(p))=0.1359>0

Note: The considered LTS in Table 2 is S={s3,s2,s1,s0,s1,s2,s3}.

(1) According to Zhang’s distance (Zhang et al.2016) (i.e. Eq. (11)), the distance between two PLTSs is zero if these two PLTSs have the same possible LTs regardless of their probabilities. Hence, such PLTSs, like L¯1(p) and L¯2(p) in Table 2, cannot be distinguished even if they are not remarkably identical.

(2) From Eq. (10), it is noticed that Pang’s distance (Pang et al.2016) has some limitations: the distance between any two PLTSs is zero if these PLTSs satisfy p1(k)r1(k)=p2(k)r2(k) (k=1,2,,#L¯(P)) for all possible LTs. In fact, PLTSs satisfying these conditions (for example, L¯3(p) and L¯4(p)) are different.

(3) Although Mao’s distance (Mao et al.2019) (i.e. Eq. (12)) improved Pang’s distance (Pang et al.2016) by replacing the subscripts ri(k) with corresponding linguistic scale functions g(L¯i(k)) (i=1,2; k=1,2,,#L¯(P)), a limitation still cannot be overcome that the distance between such PLTSs satisfying p1(k)g(L¯1(k))=p2(k)g(L¯2(k))(k=1,2,,#L¯(P)), like L¯5(p) and L¯6(p), is zero.

(4) The distance measures in this paper are developed by considering each possible element of a PLTS L¯(p) as a two-dimensional vector (i.e. L¯(k)(p¯(k))=(g(L¯(k)),p¯(k))) and then extending the corresponding classical distance measures into the PLTS context. Furthermore, it is worth noticing that the proposed distances take the hesitancy index of a PLTS into consideration. Thus, the distances proposed in this paper overcome above limitations and strengthen the distinguishing power.

3.3A Ranking Approach for PLTSs

Based on the developed distance measures, this subsection introduces a TOPSIS based ranking approach to comparing PLTSs.

Let L(p)+={sτ(1)} be a positive ideal PLTS and L(p)={sτ(1)} be a negative ideal PLTS. According to the TOPSIS method, the closer a PLTS L(p)={L(k)(p(k))|k=1,2,,#L(p)} is to L(p)+ and at the same time the farther L(p) to L(p), the better the PLTS L(p). Thus, a closeness degree of the PLTS L(p) is defined as

(16)
T(L(p))=dM(L¯(p),L¯(p))dM(L¯(p),L¯(p)+)+dM(L¯(p),L¯(p)),
where L¯(p), L¯(p)+ and L¯(p) are the normalized ordered PLTSs corresponding to L(p), L(p)+ and L(p), respectively. dM(L¯(p),L¯(p)+) and dM(L¯(p),L¯(p)) are the Manhattan distances of L¯(p) from L¯(p)+ and L¯(p), respectively.

In virtue of the closeness degree, a ranking approach for PLTSs is introduced.

Definition 14.

Let L1(p) and L2(p) be two PLTSs, where L1(p)={L1(k)(p1(k))|k=1,2,,#L1(p)} and L2(p)={L2(δ)(p2(δ))|δ=1,2,,#L2(p)}. Thus, it is stipulated as

  • (i) If T(L1(p))>T(L2(p)), then L1(p) is bigger than L2(p), denoted by L1(p)L2(p);

  • (ii) If T(L1(p))=T(L2(p)), then L1(p) is indifferent to L2(p), denoted by L1(p)L2(p);

  • (iii) If T(L1(p))<T(L2(p)), then L1(p) is smaller to L2(p), denoted by L1(p)<L2(p).

4Entropy and Cross Entropy of PLTSs

In order to judge the quality of decision information provided by DMs and the discriminations between DMs in the sequel, this section develops some entropy measures and a cross entropy measure.

4.1Entropy measures of PLTSs

This subsection addresses the entropy of PLTSs and proposes some new entropy measures of PLTSs. The main motivations are outlined as follows: (1) Measure the uncertainty of a PLTS neatly. Although Lin et al. (2019) defined an entropy measure of a PLTS, the distinguishing power of this measure is not high and cannot neatly measure the uncertainty of a PLTS. Therefore, it is necessary to develop some new entropy measures of a PLTS. (2) Overcome the shortcomings of entropy measures proposed by Liu et al. (2018). As Liu et al. (2018) stated that, the uncertainty of a PLTS includes the fuzziness and the hesitancy. To measure such fuzziness and hesitancy, Liu et al. (2018) defined a fuzzy entropy measure and a hesitancy entropy measure of a PLTS respectively. However, these entropy measures have two shortcomings: one is that some properties of the fuzzy entropy measure are not consistent with intuition. The other is that the computation of the hesitancy entropy is time-consuming. To overcome these shortcomings, this subsection proposes a new fuzzy entropy measure with desirable properties and a hesitancy entropy measure with simple computation, respectively. Finally, the total entropy of a PLTS is defined by combining the proposed fuzzy entropy and the hesitancy entropy.

4.1.1The Fuzzy Entropy of PLTSs

Definition 15

Definition 15(See Liu et al.2018).

Let S={s0,,sg/2,,sg} be a linguistic term set. Given a PLTS L(p)={li(pi)|liS;i=1,2,,#L(p)}, a fuzzy entropy of the PLTS E¯F should satisfy the following properties:

  • (i) E¯F(s0(1))=E¯F(sg(1))=0 further E¯F(s0(p),sg(1p))=0;

  • (ii) E¯F(sg/2(1))=1;

  • (iii) E¯F(L(p)(1))E¯F(L(p)(2)) if li(1)li(2)sg/2 or li(1)li(2)sg/2 and P(1)=P(2), #L(P)(1)=#L(P)(2), i=1,2,,#L(P)(1).

  • (iv) E¯F(L(p))=E¯F(L(p)(c)).

According to Definition 15, some properties of the fuzzy entropy E¯F are not in accordance with intuitions (please see Remark 4). Therefore, we define a new fuzzy entropy of a PLTS with desirable properties as below.

Definition 16.

Let S={sτ,,s0,s1,,sτ} be a linguistic term set. L¯(p), L¯1(p) and L¯2(p) are normalized ordered PLTSs, where L¯(p)={L¯(k)(p¯(k))|L¯(k)S,p¯(k)0,k=1,2,,#L¯(p)} and L¯t(p)={L¯t(k)(p¯t(k))|L¯t(k)S,p¯t(k)0,t=1,2;k=1,2,,#L¯(p),#L¯(p)=#L¯1(p)=#L¯2(p)}. We call EF a fuzzy entropy of the PLTS if it satisfies the following properties:

  • (i) EF(L¯(p))=0L¯(p)={sτ(1)}orL¯(p)={sτ(1)};

  • (ii) EF(L¯(p))=1e(L¯(p))=0.5, where e(L¯(p))=k=1#L¯(p)g(L¯(k))p¯(k) is the expected function;

  • (iii) EF(L¯1(p))EF(L¯2(p)) if L¯1(p) and L¯2(p) satisfy e(L¯1(p))e(L¯2(p))0.5 or e(L¯1(p))e(L¯2(p))0.5;

  • (iv) EF(L¯(p)(c))=EF(L¯(p)).

According to these properties, a general fuzzy entropy of PLTSs is defined below.

Theorem 4.

Suppose f:[0,1][0,1] is a strictly concave function and satisfies the following conditions:

  • (i) f(t)=f(1t) for any t[0,1];

  • (ii) f(1)=f(0)=0;

  • (iii) f is monotone increasing in [0,0.5] and monotone decreasing in [0.5,1.0]. Then the function

    (17)
    EF(L¯(p))=f(e(L¯(p)))
    is a fuzzy entropy of the PLTS L¯(p).

Proof.

Please see Appendix B.  □

Remark 4.

Compared with the fuzzy entropy E¯F, the proposed fuzzy entropy EF has the following merits:

(1) The properties of the latter are more consistent with intuitions. Let’s compare the property (i) of E¯F with that of EF. From (i) in Definition 15, we have E¯F(s0(p),sg(1p))=0. Specially, p=0.5, the equality E¯F(s0(0.5),sg(0.5))=0 means that the fuzzy uncertainty of the PLTS L(P)={s0(0.5),sg(0.5)} is zero. However, in real decision making, the decision information represented by L(P) is very fuzzy because L(P) indicates that the probability of DM dissatisfying an alternative totally and that of DM satisfying an alternative completely are both 0.5. Therefore, the property E¯F(s0(0.5),sg(0.5))=0 is counter-intuitive. However, according to the proposed fuzzy entropy EF defined in Eq. (17), one gets E¯F(s0(p),sg(1p))=f(p). This indicates that the fuzzy entropy of L(P) varies with p. Employing Theorem 4, we have EF(s0(0.5),sg(0.5))=f(0.5)=1. At this point, the proposed entropy EF is more consistent with intuition.

(2) The latter can be regarded as an extension of the former in some sense. It is worth mentioning that a normalized PLTS can be considered as a type of discrete random variables. Therefore, the expectation value E(L¯(P)) represents the average level of L¯(P). Based on this idea, it is reasonable that the decision information described by L¯(P) and that of {sg/2(1)} in Definition 15 are equivalent if E(L¯(P))=0.5. Hence, the property E¯F(sg/2(1))=1 in Definition 15 can be extended to the one EF(L¯(p))=1e(L¯(p))=0.5 in Definition 16. Likewise, the property (iii) in Definition 15 is also extended into the one (iii) in Definition 16.

According to Theorem 4, let f(t)=ln(1+tt2)ln5ln4. A fuzzy entropy of a PLTS is defined as

(18)
EF(L¯(p))=ln(1+e(L¯(p))(e(L¯(p)))2)ln5ln4.

4.1.2The Hesitancy Entropy of PLTSs

Hesitancy is an important feature of a PLTS. How to measure the hesitant uncertainty of a PLTS is often ignored. To fill in this gap, this subsection defines a hesitancy entropy measure of PLTSs by the deviations between the linguistic scale function values of possible linguistic terms in a PLTS and their mean value.

Definition 17.

Let S={sτ,,s0,s1,,sτ} be a linguistic term set. L¯(p), L¯1(p) and L¯2(p) are normalized ordered PLTSs, where L¯(p)={L(k)(p¯(k))|L(k)S,p¯(k)0,k=1,2,,#L¯(p)} and L¯t(p)={Lt(k)(p¯t(k))|Lt(k)S,p¯t(k)0,t=1,2;k=1,2,,#L¯(p),#L¯(p)=#L¯1(p)=#L¯2(p)}. We call EH a hesitancy entropy of a PLTS if it satisfies the following properties:

  • (i) EH(L¯(p))=0L¯(p)={sα(1)|α=τ,,0,1,,τ};

  • (ii) EH(L¯(p))=1L¯(p)={sτ(0.5),sτ(0.5)};

  • (iii) EH(L¯1(p))0 if L¯(p)={L(1)(p),L(2)(1p)} and L(1)L(2).

  • (iv) EH(L¯1(p))EH(L¯2(p)) if h(L¯1(p))h(L¯2(p)) for any i,j=1,2,,#L¯(p).

  • (v) EH((L¯(p))(c))=EH(L¯(p)).

Theorem 5.

Given a normalized ordered PLTS L¯(p). The hesitancy index

(19)
Eh(L¯(p))=h(L¯(p))=2i=1#L¯(p)|g(L(i))g¯|pi#L¯(p)>1,0#L¯(p)=1,
is a hesitancy entropy, where g¯=1#L¯(p)i=1#L¯(p)g(L(i)).

Proof.

Please see Appendix C.  □

Remark 5.

Let S={s0,,sg/2,,sg} be a linguistic term set. Given a PLTS L(p)={li(pi)|liS;i=1,2,,#L(p)}, Liu et al. (2018) defined a hesitancy entropy measure as follows:

(20)
E¯h(L¯(p))=i=1#L¯(p)1j=i+1#L¯(p)4pipjf(γij)#L¯(p)>1,0#L¯(p)=1,
where γij=|I(li)I(lj)|g and I(li) is the subscript of the linguistic term li.

Compared E¯h with Eh, the difference between them is that the former measures the deviations between different linguistic terms in a PLTS, while the latter measures the deviations between all the transformation function values of possible linguistic terms and their mean value. Although the expressions of E¯h and Eh are different, they have similar properties (please see Definition 5 in Liu et al. (2018). However, the computation of Eh needs #L¯(P)+1 times, while E¯h needs #L¯(P)(#L¯(P)1)/2 times. Thus, the computations of Eh is more time-saving when #L¯(P)4.

Combining the fuzzy entropy and hesitancy entropy with an adjusted coefficient θ, a total entropy is determined as

(21)
ET(L¯(p))=θEF(L¯(p))+(1θ)EH(L¯(p)).

As 0EF(L¯(p)), EH(L¯(p))1, one has 0ET(L¯(p))1.

Furthermore, the total entropy of a probabilistic linguistic matrix is derived.

Definition 18.

Let U¯=(L¯ij(p))m×n be a matrix with PLTSs. The total entropy of U¯ is defined as

(22)
ET(U¯)=i=1mj=1nET((L¯ij(p)).

Remark 6.

Lin et al. (2019) proposed an information entropy as

(23)
μ(L(p))=k=1#L(p)p(k)log2p(k),
where L(p) is a PLTS, #L(p) is the number of all different linguistic terms in L(p) and p(k) is the probability of kth possible linguistic value of L(p).

It is obvious from Eq. (23) that this information entropy only considered the probabilities of possible linguistic terms in a PLTS, but ignored all possible linguistic terms. However, the total entropy presented in this paper takes them into consideration together. Therefore, compared with μ(L(p)) in Eq. (23), the presented total entropy ET(L¯(p)) has a stronger distinguishing power while measuring the uncertainty of a PLTS, which can be verified by Example 2.

Example 2.

Given two PLTSs L1(p)={s3(0.30),s0(0.25),s3(0.45)} and L2(p)={s1(0.30),s2(0.25),s3(0.45)}. Obviously, the uncertainty of L1(p) is larger than that of L2(p). In fact, the total entropy is calculated as ET(L1(p))=0.552>ET(L2(p))=0.398, which is in accordance with our intuition. However, by Eq. (23), it obtains that μ(L1(p))=μ(L2(p))=1.5395, by which the uncertainties of L1(p) and L2(p) are considered the same. In other words, μ(L(p)) in Eq. (23) is unable to distinguish the degrees of uncertainty of L1(p) and L2(p). Therefore, the total entropy proposed in this paper has a stronger distinguishing power.

4.2The Cross Entropy of PLTSs

In this subsection, to measure the discrimination degree between PLTSs, a cross entropy is defined. Motivated by the cross entropy in the hesitant fuzzy linguistic environment (Gou et al.2017), the axiomatic definition of cross-entropy measure for PLTSs is given below.

Definition 19.

Suppose S={sτ,,s0,s1,,sτ} is a linguistic term set. L¯1(p) and L¯2(p) are two normalized ordered PLTSs with #L¯1(p)=#L¯2(p), where L¯t(p)={L¯t(k)(p¯t(k))|L¯t(k)S,p¯t(k)0,t=1,2;k=1,2,,#L¯1(p)}. Then the cross entropy CE(L¯1(p),L¯2(p)) of L¯1(p) and L¯2(p) should satisfy the following conditions:

  • (i) CE(L¯1(p),L¯2(p))0;

  • (ii) CE(L¯1(p),L¯2(p))=0 if and only if e(L¯1(p))=e(L¯2(p)).

Theorem 6.

Given two normalized ordered PLTSs L¯1(p) and L¯2(p), the function

(24)
CE(L¯1(p),L¯2(p))=e(L¯1(p))ln2e(L¯1(p))e(L¯1(p))+e(L¯2(p))+(1e(L¯1(p)))ln2(1e(L¯1(p)))1e(L¯1(p))+1e(L¯2(p))
is a cross entropy.

Proof.

Please see Appendix D.  □

It is observed from Eq. (24) that the cross entropy, CE(L¯1(p),L¯2(p)), is not symmetric. A symmetric cross entropy between PLTSs can be obtained as

(25)
D(L¯1(p),L¯2(p))=CE(L¯1(p),L¯2(p))+CE(L¯2(p),L¯1(p)).

Moreover, a symmetric cross entropy between two probabilistic linguistic matrices is defined below.

Definition 20.

Let U¯1=(L¯ij1(p))m×n and U¯2=(L¯ij2(p))m×n be two matrices with PLTSs. The symmetrical cross entropy between U¯1 and U¯2 is defined as

(26)
D(U¯1,U¯2)=i=1mj=1nD(L¯ij1(p),L¯ij2(p)).

5A Novel Method for MAGDM with PLTSs

In this section, we first describe the problems of MAGDM with PLTSs, and then propose a novel method for solving such problems.

5.1Problem Description

Let A={A1,A2,,Am} be the set of m feasible alternatives, U={u1,u2,,un} be the set of attributes whose weights are W=(w1,w2,,wn)T with j=1nwj=1, and E={e1,e2,,et} be the set of DMs whose weights are λ=(λ1,λ2,,λt)T satisfying k=1tλk=1. DM ek (k=1,2,,t) provides his/her evaluations on alternative Ai with respect to attribute uj is in the form of PLTS Lijk(p) (i=1, 2,,m;j=1, 2,,n;k=1,2,,t) based on the linguistic term set S={sα|α=τ,,1,0,1,,τ}. All evaluations construct the decision matrices Uk=(Lijk(p))m×n (k=1,2,,t). Denote the normalized ordered decision matrices by U¯k=(L¯ijk(p))m×n (k=1,2,,t).

5.2Determine DMs’ Weights Based on the Total Entropy and the Symmetric Cross Entropy

In this subsection, an approach is developed to determine DMs’ weights objectively by using the proposed total entropy and symmetric cross entropy of PLTSs. As we know, the less uncertainty (i.e. a smaller total entropy) of an individual matrix provided by a DM, the better the quality of decision information reflected by this matrix is. Thus, the bigger the weight of this DM should be assigned. In virtue of this criterion, a programming model for deriving DMs’ weights is built by minimizing the total entropy of decision matrices, i.e.

(27)
(M1)minF(λ1k)=1tk=1tET(U¯k)λ1k,s.t.k=1t(λ1k)2=1,
where ET(U¯k) are the total entropy of U¯k (k=1,2,,t).

Solving Eq. (27) with Lagrange multiplier approach, it is obtained as

(28)
λ1k=ET(U¯k)k=1t(ET(U¯k))2.

Normalizing λ1k, one has

(29)
λ1k=ET(U¯k)k=1tET(U¯k)(k=1,2,,t).

On the other hand, a closer degree of a DM to other DMs means that the information supplied by this DM is much closer to that of the group. In this case, this DM should be assigned a larger weight. From this viewpoint, another optimal model for determining DMs’ weights is constructed based on the symmetric cross entropy as

(30)
(M2)minF(λ2k)=1t1k=1tδ=1,δktD(U¯k,U¯δ)λ2k,s.t.k=1t(λ2k)2=1.

Eq. (30) can also be solved with Lagrange multiplier approach, and the optimal solutions are derived as

(31)
λ2k=1t1δ=1,δktD(U¯k,U¯δ)k=1t(1t1δ=1,δktD(U¯k,U¯δ))2(k=1,2,,t).

Normalizing λ2k, one gets

(32)
λ2k=δ=1,δktD(U¯k,U¯δ)k=1tδ=1,δktD(U¯k,U¯δ)(k=1,2,,t).

Combining Eq. (29) and Eq. (32), the ultimate DMs’ weights λk are determined as

(33)
λk=βλ1k+(1β)λ2k(k=1,2,,t),
where β[0,1] is a compromise coefficient.

5.3Construct Bi-Objective Programs For Deriving Attribute Weights

In group decision making process, attribute weights play an important role because different attribute weights may result in diverse ranking of alternatives. In this subsection, considering the information of attribute weights is completely unknown or partially known, two bi-objective programs are constructed respectively for deriving attribute weights.

(i) Aggregating individual decision matrices into a collective one

By employing DMs’ weights determined in Section 5.2, individual decision matrices are integrated into a collective one with the proposed PLWAM operator. Denote the collective decision matrix by U=(Lij(p))m×n, where

(34)
Lij(p)=PLWAM(Lij1(p),Lij2(p),,Lijn(p)).
Then, the normalized ordered collective decision matrix U¯=(L¯ij(p))m×n is obtained from U=(Lij(p))m×n by Definition 4, where L¯ij(p)={L¯ij(k)(p¯ij(k))|L¯ij(k)S,p¯ij(k)0,k=1,2,,#L¯ij(p)}.

(ii) Constructing bi-objective programs for deriving attribute weights

It is known to us that an attribute plays a more important role if the performance values of alternatives on it have distinct differences. Thus, such an attribute should be given a bigger weight. Conversely, if the evaluation values of alternatives with respect to a certain attribute have little difference, this attribute should be given a smaller weight. In addition, the credibility of decision information on an attribute should also be taken into account while determining attribute weights. The more credibility of evaluation values on an attribute, the bigger weight of this attribute should be assessed. As mentioned before, the symmetric cross entropy and the total entropy can reflect the differences between alternatives and the credibility of evaluation values, respectively. Keeping this idea in mind, we can determine attribute weights by maximizing the symmetric cross entropy as well as minimizing the total entropy. Thus, a bi-objective program is established if the information of attribute weights is completely unknown, i.e.

(35)
(M3)maxF(w)=j=1mFj(w)=j=1m(i=1nr=1,rinD(L¯ij(p),L¯rj(p)))wj,minT(w)=j=1mTj(w)=j=1m(i=1nET(L¯ij(p))wj),s.t.k=1t(wj)2=1.

As 0ET(L¯ij(p))1, Eq. (35) can be converted into the following single objective program:

(36)
(M4)maxF(w)=j=1mFj(w)=j=1mi=1n(r=1,rinD(L¯ij(p),L¯rj(p))+(1ET(L¯ij(p))))wj,s.t.k=1t(wj)2=1.

Solving Eq. (36) with Lagrange multiplier approach, the weights of attributes are derived as

(37)
wj=i=1n(r=1,rinD(L¯ij(p),L¯rj(p))+(1ET(L¯ij(p))))j=1m(i=1n(r=1,rinD(L¯ij(p),L¯rj(p)))+(1ET(L¯ij(p)))))2(j=1,2,,n).

The normalized weights wj are obtained by normalizing wj, i.e.

(38)
wj=i=1n(r=1,rinD(L¯ij(p),L¯rj(p))+(1ET(L¯ij(p))))j=1m(i=1n(r=1,rinD(L¯ij(p),L¯rj(p))+(1ETL¯ij(p))))(j=1,2,,n).

Similarly, for the situations where the information of attribute weights is incomplete, another programming model is obtained by modifying Eq. (36), i.e.

(39)
(M5)maxF(w)=j=1mFj(w)=j=1mi=1n(1n1r=1,rinD(L¯ij(p),L¯rj(p))+(1ET(L¯ij(p))))wj,s.t.wjΩ;k=1t(wj)2=1,
where Ω represents the incomplete attribute weight information. Solving Eq. (39) with Lingo software, the corresponding optimal solution vector w=(w1,w2,,w2)T is generated and then normalized as the attribute weight vector w=(w1,w2,,wn)T.

5.4Ranking Alternatives by an Improved PL-PROMETHEE Method

The PROMETHEE method is a popular outranking method in decision making. It has been extended to different fuzzy environments, such as intuitionistic fuzzy sets (Krishankumar et al.2017) and linguistic variables Halouani et al. (2009). In the PLTS environment, although Xu et al. (2019) proposed a PL-PROMETHEE method, the distinguishing power of PL-PROMETHEE is not high, which can be verified by Example 3 in the sequel. To improve the distinguishing power, this subsection constructs new probabilistic linguistic preference relations, the integrated preference index and positive/negative outranking flows to adapt the main framework of PROMETHEE in the PLTS context. Thus, an improved PL-PROMETHEE method is proposed to solve MAGDM problems with PLTSs.

(i) The probabilistic linguistic preference function

Suppose that Ai and Ar are two alternatives in the alternative set. L¯ij(p) and L¯rj(p) are respectively normalized ordered attribute values of Ai and Ar with respect to attribute uj. Preference function PXj(Ai,Ar) of L¯ij(p) with respect to L¯rj(p) is defined as

(40)
PXj(Ai,Ar)=Ljir(p)=k1=1,2,,#L¯ij(p)k2=1,2,,#L¯rj(p){P(Lij(k1),Lrj(k2))}(ir),
where
(41)
P(Lij(k1),Lrj(k2))=sτ(p¯ij(k1)p¯rj(k2))ifg(L¯ij(k1))p¯ij(k1)g(L¯rj(k2))p¯rj(k2)q,g1(g(L¯ij(k1))p¯ij(k1)g(L¯rj(k2))p¯rj(k2)p)(p¯ij(k1)p¯rj(k2))ifq<g(L¯ij(k1))p¯ij(k1)g(L¯rj(k2))p¯rj(k2)p,sτ(p¯ij(k1)p¯rj(k2))ifg(L¯ij(k1))p¯ij(k1)g(L¯rj(k2))p¯rj(k2)>p.

In Eq. (41), parameters p(0) and q(0) are the indifference threshold and the strict preference threshold, respectively. Functions g(x) and g1(x) are defined in Definition 2.

Obviously, the preference function PXj(Ai,Ar) constructed by Eqs. (40) and (41) is a PLTS.

(ii) The integrated preference index

According to the probabilistic linguistic preference index PXj(Ai,Ar), an integrated preference index is defined as

(42)
πX(Ai,Ar)=j=1nwjT(PXj(Ai,Ar)),
where T(Pj(Ai,Ar)) is the closeness degrees of PXj(Ai,Ar) and computed by Eq. (16).

(iii) The positive and negative outranking flows

Employing the integrated preference indices, the positive/negative outranking flows, ϕ+(Ai) and ϕ(Ai), are defined as follows.

(43)
ϕ+(Ai)=1m1r=1,rimπX(Ai,Ar).
(44)
ϕ(Ai)=1m1r=1,rimπX(Ar,Ai).

(iv) Ranking alternatives.

In virtue of Eqs. (43) and (44), the net flow of alternative Ai is calculated as

(45)
ϕ(Ai)=ϕ+(Ai)ϕ(Ai).

Finally, alternatives are ranked by the descending order of net flows.

Remark 7.

In PL-PROMETHEE method (Xu et al.2019), the preference function and the total preference index are defined respectively as

(46)
Pj(Ai,Ar)=0,p(Ai,j>Ar,j)0.5,1,p(Ai,j>Ar,j)>0.5
and
(47)
π(Ai,Ar)=j=1mPj(Ai,Ar)wj,
where P(Ai,j>Ar,j) is the probability of Ai preferred to Ar with respect to uj.

Obviously, the value of the preference function Pj(Ai,Ar) in Eq. (46) only takes 0 or 1, which may result in the loss of decision information. However, the preference function PXj(Ai,Ar) in Eq. (40) is described by the PLTS which is more flexible than the crisp number for representing decision information. Therefore, compared with PL-PROMETHEE, the distinguishing power of the improved PL-PROMETHEE is stronger and the sensitivity to the ratings of alternatives on attributes is increased. To illustrate this advantage, Example 3 is given below.

Example 3.

For convenience, we only consider one of attributes in a real problem. Suppose that the ratings of three alternatives on this attribute are L1(p)={s2(0.35),s3(0.25),s4(0.40)}, L2(p)={s3(0.70),s2(0.30)} and L3(p)={s1(0.40),s2(0.25),s4(0.35)}, respectively. Intuitively, the relation L1(p)L3(p)L2(p) holds, where the symbol “≻” means “preferred to”. Hence, it is deduced that the degree of L1(p) preferred to L2(p) should be more than that of L1(p) preferred to L3(p). In fact, take q=0 and p=0.5 in the proposed preference function (i.e. Eq (41)), the preference values are calculated as PX1(A1,A2)={s1.9(0.175),s1.7(0.075),s1.2(0.245),s1(0.105),s1(0.280),s1.2(0.120)} and PX1(A1,A3)={s4(0.31),s3.8(0.140),s3.5(0.063),s3.2(0.14),s2.8(0.088),s1.6(0.160),s0.6(0.100)}. The preference index is calculated as πX(A1,A2)=0.455>πX(A1,A3)=0.306. This result is in line with human intuition. However, using the preference function and preference index in method (Xu et al.2019) (i.e. Eqs. (46) and (47)), one obtains P1(A1,A2)=P1(A1,A3)=1 and π12=π13=1, which are not consistent with the above analysis. In other words, method (Xu et al.2019) has no ability to distinguish alternatives A2 and A3. Therefore, the improved PL-PROMETHEE method has a stronger distinguishing power than the method (Xu et al.2019). Furthermore, the sensitivity of the improved PL-PROMETHEE method is strong, too. For example, when L2(p) is increased to L2(p)={s2(0.80),s3(0.20)} while L1(p) and L3(p) remain unchanged, the corresponding preference index πX(A1,A2) decreases to 0.271 from 0.455, whereas π12 is not changed.

5.5A Novel Method for MAGDM with Probabilistic Linguistic Information

A novel method is generated for MAGDM problems with PLTSs. The main procedure of this method is outlined below.

Step 1. Each DM establishes his/her individual probabilistic linguistic matrices Uk=(Lijk(p))m×n (k=1,2,,t). Further, matrices Uk are transformed into corresponding normalized ordered matrices U¯k=(L¯ijk(p))m×n (k=1,2,,t) by Definitions 4-6.

Step 2. Calculate the total entropy ET(U¯k) (k=1,2,,t) and the symmetric cross entropy D(U¯k,U¯δ) (k,δ=1,2,,t;kδ) by Eqs. (18), (19), (21), (24) and (25).

Step 3. Determine the weight vector of DMs λ=(λ1,λ2,,λt)T by Eqs. (29), (32) and (33).

Step 4. Aggregate all matrices U¯k=(Lijk(p))m×n (k=1,2,,t) into a collective one U=(Lij(p))m×n by the PLWAM operator (Eq. (4)), and then transform matrix U into a normalized ordered matrix U¯=(L¯ij(p))m×n.

Step 5. Calculate the entropy ET(L¯ij(p)) (i=1, 2,,m;j=1,2,,n) and the symmetric cross entropy D(L¯ij(p),L¯rj(p)) (i,r=1,2,,m; ri; j=1,2,,n) by Eqs. (18), (19), (21), (24) and (25).

Step 6. Determine the weight vector of attributes w=(w1,w2,,wn)T by Eq. (38) or (39).

Step 7. Compute the preference function values PXj(Ai,Ar) (i,r=1,2,,m; ri; j=1,2,,n) by Eqs. (40) and (41).

Step 8. Obtain the total preference index matrix π=(πX(Ai,Ar))m×m by Eq. (42).

Step 9. Calculate the positive and negative flows ϕ+(Ai) and ϕ(Ai) (i=1,2,,m) based on Eqs. (43) and (44).

Step 10. Determine net flows ϕ(Ai) (i=1,2,,m) by Eq. (45). Alternatives are ranked based on the descending orders of ϕ(Ai) (i=1,2,,m).

6A Case Study

In this section, an example of a car sharing platform selection is provided to illustrate the application of the proposed method. Furthermore, the comparative analyses are performed to show the merits of the proposed method.

6.1Car Sharing Platform Selection

With the rapid development of internet technology and the deep advocation of green travel, the car sharing has sprung up over the last three years. Up to now, several car sharing platforms have emerged in China, such as Evcard, Gofun, Togo and so on. The popularization of car sharing greatly facilitates peoples’ travel and relieves the traffic pressure.

As a famous tourist city in China, Guilin cannot satisfy the travel of tourists due to the limited operational capacity of public transportation. So it is necessary for the government to introduce a car sharing platform to resolve the traffic problem. Now, the government invites three DMs to select the best car sharing platform from four candidate platforms (alternatives), including Evcard, Gofun, Togo and Greengo. Four attributes are considered, including safety (u1), convenience (u2), service (u3) and car brand (u4). Suppose a linguistic term set S={s4,s3,s2,s1,s0,s1,s2,s3,s4} is provided in Table 3. Three invited DMs, d1, d2 and d3, utilize PLTSs to evaluate four alternatives with respect to four attributes, and construct three probabilistic linguistic decision matrices U1, U2 and U3.

Step 1. Each DM establishes decision matrices U1, U2 and U3 as shown in Table 4, and the corresponding normalized ordered matrices U¯1, U¯2 and U¯3 are listed in Table 5.

Table 3

Linguistic variables corresponding to linguistic terms.

Linguistic variablesLinguistic termsLinguistic variablesLinguistic terms
Very bads4Slightly goods1
Bads3A little goods2
A little bads2Goods3
Slightly bads1Very goods4
Mediums0
Table 4

Probabilistic linguistic decision matrices U1, U2 and U3.

u1u2u3u4
d1A1{s2(0.4),s1(0.5)}{s2(0.6),s4(0.4)}{s0(1)}{s2(0.4),s1(0.6)}
A2{s4(1)}{s2(0.4),s4(0.5)}{s0(0.3),s1(0.3),s2(0.4)}{s2(0.3),s3(0.7)}
A3{s1(0.3),s2(0.7)}{s1(0.4),s2(0.3),s3(0.3)}{s2(1)}{s1(0.8),s1(0.2)}
A4{s1(1)}{s1(0.6),s2(0.4)}{s0(0.5),s2(0.5)}{s3(1)}
d2A1{s1(0.3),s1(0.7)}{s2(0.4),s3(0.2),s4(0.4)}{s0(0.7),s1(0.3)}{s1(1)}
A2{s3(0.4),s4(0.5)}{s3(1)}{s3(0.5),s4(0.5)}{s0(0.4),s2(0.6)}
A3{s0(0.7),s1(0.3)}{s1(0.6),s2(0.3)}{s1(0.3),s2(0.7)}{s1(1)}
A4{s2(1)}{s1(0.5),s2(0.4)}{s1(1)}{s1(0.1),s2(0.2),s3(0.7)}
d3A1{s0(0.4),s1(0.6)}{s3(0.6),s4(0.4)}{s1(0.2),s0(0.8)}{s1(1)}
A2{s3(0.4),s4(0.6)}{s1(0.3),s2(0.7)}{s3(0.6),s4(0.4)}{s2(1)}
A3{s0(0.6),s1(0.4)}{s1(1)}{s1(0.2),s2(0.8)}{s1(0.7),s1(0.3)}
A4{s4(0.4),s2(0.4)}{s1(0.5),s2(0.5)}{s1(0.8),s0(0.2)}{s2(1)}

Step 2. Calculate the total entropy and the symmetric cross entropy.

Table 5

Normalized ordered decision matrices U¯1, U¯2 and U¯3.

u1u2u3u4
d1A1{s2(4/9),s2(0),s1(5/9)}{s2(0),s2(0.6),s4(0.4)}{s0(0),s0(0),s0(1)}{s2(0.4),s1(0.6),s2(0)}
A2{s4(0),s4(0),s4(1)}{s2(0),s2(4/9),s4(5/9)}{s0(0.3),s1(0.3),s2(0.4)}{s2(0),s2(0.3),s3(0.7)}
A3{s1(0),s1(0.3),s2(0.7)}{s1(0.4),s2(0.3),s3(0.3)}{s2(0),s2(0),s2(1)}{s1(0.8),s1(0),s1(0.2)}
A4{s1(1),s1(0),s1(0)}{s1(0),s1(0.6),s2(0.4)}{s0(0),s0(0.5),s2(0.5)}{s3(0),s3(0),s3(1)}
d2A1{s1(0.3),s1(0),s1(0.7)}{s3(0.2),s2(0.4),s4(0.4)}{s0(0),s0(0.7),s1(0.3)}{s1(1),s1(0),s1(0)}
A2{s3(0.),s3(4/9),s4(5/9)}{s3(0),s3(0),s3(1)}{s3(0),s3(0.5),s4(0.5)}{s0(0),s0(0.4),s2(0.6)}
A3{s0(0),s0(0.7),s1(0.3)}{s1(0),s1(6/9),s2(3/9)}{s1(0),s1(0.3),s2(0.7)}{s1(1),s1(0),s1(0)}
A4{s2(1),s2(0),s2(0)}{s1(0),s1(5/9),s2(4/9)}{s1(1),s1(0),s1(0)}{s1(0.1),s2(0.2),s3(0.7)}
d3A1{s0(0),s0(0.4),s1(0.6)}{s3(0),s3(0.6),s4(0.4)}{s1(0.2),s1(0),s0(0.8)}{s1(1),s1(0),s1(0)}
A2{s3(0),s3(0.4),s4(0.6)}{s1(0),s1(0.3),s2(0.7)}{s3(0),s4(0.4),s3(0.6)}{s0(0),s0(0.1),s1(0.9)}
A3{s0(0),s0(0.6),s1(0.4)}{s1(0),s1(0),s1(1)}{s1(0.2),s1(0),s2(0.8)}{s1(0.7),s1(0),s1(0.3)}
A4{s4(0.5),s2(0.5),s4(0)}{s1(0),s1(0.5),s2(0.5)}{s1(0.8),s1(0),s0(0.2)}{s2(0),s2(0),s2(1)}

By Eqs. (18), (19), (21) and (22), the total entropies of matrices U¯1, U¯2 and U¯3 are respectively calculated as ET(U¯1)=7.4022, ET(U¯2)=7.4223, ET(U¯2)=7.7322. Employing Eqs. (24)–(26), the symmetric cross entropies are obtained and represented in Table 6.

Step 3. Determine the weight vector of DMs.

Table 6

The symmetric cross entropies between U¯1, U¯2 and U¯3.

U¯i U¯j D(U¯k, U¯δ)U¯1U¯2U¯3
U¯10.28040.3965
U¯20.28040.084
U¯30.39650.084

Suppose β=0.15, and the weight vector of DMs is derived by Eqs. (29), (32) and (33) as λ=(0.4273,0.2529,0.3198)T.

Step 4. Aggregate individual decision matrices into a collective one.

By employing the PLWAM operator (see Eq. (4)), individual normalized ordered matrices U¯k (k=1,2,3) are aggregated into a collective one U=(Lij(p))m×n, which is converted into an normalized ordered matrix U¯=(L¯ij(p))m×n, please see Table 7.

Steps 5. Calculate the total entropy and symmetric cross entropy of alternatives.

Table 7

Collective normalized ordered matrix U¯.

u1u2u3u4
d1A1{s1.03(0.05),s0.59(0.08),s0.42(0.12),s0.03(0.19),s0.26(0.07),s0.58(0.10),s0.71(0.16),s1(0.23)}{s2.40(0.14),s2.66(0.07),s4(0.78)}{s0.30(0.14),s0(0.56),s0.01(0.06),s0.28(0.24)}{s1.40(0.4),s1(0.6)}
A2{s4(1)}{s2.10(0.13),s2.32(0.31),s4(0.56)}{s2.19(0.09),s2.41(0.09),s2.66(0.12),s4(0.7)}{s1.02(0.01),s1.29(0.11),s1.50(0.00),s1.72(0.16),s1.78(0.03),s1.98(0.25),s2.14(0.04),s2.31(0.38)}
A3{s0.46(0.13),s0.71(0.05),s0.77(0.08),s1(0.04),s1.02(0.29),s1.23(0.13),s1.29(0.20),s1.48(0.08)}{s1(0.27),s1.29(1.33),s1.47(0.20),s1.72(0.10),s2.10(0.20),s2.30(0.10)}{s1.03(0.06),s1.32(0.14),s1.78(0.24),s2.00(0.56)}{s1(0.56),s0.25(0.24),s0.02(0.14),s0.59(0.06)}
A4{s2.1(0.5),s1.56(0.5)}{s1(0.17),s1.30(0.13),s1.37(0.17),s1.47(0.11),s1.63(0.13),s1.72(0.09),s1.78(0.11),s2.00(0.09)}{s0.55(0.40),s0.23(0.10),s0.61(0.40),s0.85(0.10)}{s2.35(0.1),s2.51(0.2),s2.75(0.7)}

Take θ=0.5, and the total entropy matrix E=ET(L¯ij(p)) is obtained by Eqs. (18), (19) and (21) as

E=0.58780.49900.52250.50840.41700.55380.52460.53860.53110.53430.53690.55610.47800.51540.57170.4783.

Employing Eqs. (24) and (25), the symmetric cross entropy matrix, denoted by D=(Dij)4×4, where Dij=r=1,ri4D(L¯ij(p),L¯rj(p)) (j=1,2,3,4), is derived as

D=0.33861.10160.29560.57301.10160.17670.44840.23100.29560.44840.17790.26620.57300.23100.26620.4789.

Step 6. Determine the weight vector of attributes.

The weight vector of attributes is determined by Eq. (38) as

w=(0.2932,0.2632,0.2070,0.2367)T.

Step 7. Compute the preference function values.

Using Eqs. (40) and (41), preference function values are computed. In view of limited space, the concrete preference function values are not listed.

Step 8. Obtain the integrated preference index.

By Eq. (42), the total preference index matrix is derived as

π=0.22320.50000.22390.43230.44220.42340.15410.15330.14720.14410.22180.2001.

Step 9. Calculate positive and negative flows of alternatives.

In virtue of Eqs. (43) and (44), positive and negative flows of alternatives are respectively calculated as

ϕ+(A1)=0.3157,ϕ+(A2)=0.4326,ϕ+(A3)=0.1515,ϕ+(A4)=0.1887.ϕ(A1)=0.2435,ϕ(A2)=0.1994,ϕ(A3)=0.3808,ϕ(A4)=0.2648.

Step 10. Rank alternatives.

Net flows of alternatives are obtained by Eq. (45) as

ϕ(A1)=0.0722,ϕ(A2)=0.2332,ϕ(A3)=0.2292,ϕ(A4)=0.0762.

As ϕ(A2)>ϕ(A1)>ϕ(A4)>ϕ(A3), alternatives are ranked as A2A1A4A3. Therefore, the alternative A2 (Gofun platform) is the best one.

6.2Comparative Analyses

To show the advantages of the proposed method, comparative analyses with Mao’s method (Mao et al.2019), PL-PROMETHEE method (Xu et al.2019) and other existing decision methods are performed in the sequel.

6.2.1Comparison with Mao’s method

Mao et al. (2019) presented a new method integrating ELECTRE and TOPSIS to solve MAGDM problems with PLTSs. Now we use Mao’s method (Mao et al.2019) to solve the above problem. Set γ=1 and λ=1. DMs’ subjective weight vector is assigned as λ1=(0.04,0.88,0.08)T. Thus, the corresponding ordering results are derived as A1A2A4A3. Clearly, these results are not consistent with those obtained by the proposed method, which can be explained as follows:

(1) The proposed method determines DMs’ weights objectively by the total entropy and cross entropy of decision information. However, method (Mao et al.2019) assigned DMs’ weights subjectively, by which the random cannot be avoided. Although an adjusted coefficient obtained from the consistency degree between DMs is used to adjust the subjective weights of DMs, the adjusted DMs’ weights may be irrational because the consistency degree ρir indicating the consistency between DM di and dr (i.e. Eq. (40) in method Mao et al.2019) does not satisfy the symmetry. For example, the consistency degree ρ12=0.5125, while ρ21=0.4799, So ρ21ρ12. In fact, the consistency degree should be symmetrical theoretically. Hence, DMs’ weights derived by the proposed method are more reasonable.

(2) The aggregated value obtained by the proposed PLWAM operator is more convincing than that obtained by GPLHWAw operator in Mao et al. (2019). The former is a PLTS based on Theorem 1. Conversely, the latter cannot always satisfy this property. For example, while solving example in Section 6.1, L11(p)=GPLHWAw(L¯111(p),L¯112(p),L¯113(p))={g1(0.6937)(0.442),g1(0.7449)(0.3),g1(1)(0.7691),g1(0.7521)(0.3750),g1(0.2191)}. Clearly, one has 0.442+0.3+0.7691+0.7521+0.2191=2.375>1. Hence, the latter sometimes is not a PLTS in a strict sense. Thus, the result derived by the proposed method is more convincing.

6.2.2Compared with the PL-PROMETHEE Method

The PL-PROMETHEE method (Xu et al.2019) is applied to solve the problem in Section 6.1. Suppose attribute weight vector as w=(0.3,0.2,0.3,0.2)T. The Borda’s scores of alternatives for each DM are shown as Table 8.

Table 8

The Borda’s scores of alternatives for each alternatives.

d1d2d3Borda’s scores
A11438
A24217
A33328
A42147

It finds from Table 8 that the ranking result is A1A3A2A4, which is different from the one derived by the proposed method, i.e, A2A1A4A3. The primary reasons may result from the following two aspects.

(1) PL-PROMETHEE method ignored the determination of attribute weights and DMs’ weights, but assigned them subjectively. Thus, the arbitrariness cannot be avoided. By contrast, the proposed method derives attribute weights and DMs’ weights objectively based on the total entropy and symmetric cross entropy of decision information. Therefore, the subjectivity is effectively reduced and the decision results are more credible.

(2) In the preference function of PL-PROMETHEE method (Xu et al.2019), the preference value is taken as 1 if the probability of one alternative preferred to the other with respect to an attribute is bigger than 0.5. Otherwise, the preference value is taken as 0. Therefore, the distinguishing power of this preference function is not STRONG enough, which may be the reason why alternatives A1 and A3 cannot be discriminated. However, in the proposed method, the preference function is neatly defined based on the deviations of attribute values and described by PLTSs, please see Eqs. (40) and (41). Hence, the proposed method has a stronger distinguishing power.

6.2.3Compared with Other Existing Decision Making Methods

To further demonstrate the superiority of the proposed method, this subsection conducts a theoretical analysis and a practical analysis with other existing methods (Pang et al.2016; Gou et al.2017; Liu and Li, 2018; Mao et al.2019; Liu and Li, 2019; Peng et al.2020).

(1) Theoretical analysis

(i) The proposed method introduces two new probabilistic linguistic weighted average operators (i.e. PLWAM operator and PLWAG operator). Compared with existing operators mentioned in Pang et al. (2016) and Gou et al. (2017), the proposed operators have some advantages, such as the closure of operations and operation values being consistent with intuition (please see Remark 2). Therefore, the aggregated information obtained by the proposed operators is more reliable, and thus the decision results based on such aggregated information are more reasonable.

(ii) In the proposed method, two cases including the attribute weights with unknown or partially known values are both taken into account. To determine attribute weights, the proposed method builds two different bi-objective programming models by maximizing the cross entropy and minimizing the total entropy of the collective evaluation values. As analysed in Section 5.3, this model considers the quality of the collective evaluations and the deviations between evaluations. However, in the extended MULTIMOORA method (Liu and Li, 2019), the attribute weights were given in advance, which may be not able to avoid the subjective arbitrariness in decision process. Although the TOPSIS method (Pang et al.2016) derived attribute weights objectively by the maximum deviation approach, it ignored the quality of decision information. Thus, the creditability of attribute weights derived by the proposed method is higher.

(iii) Although method (Peng et al.2020) and the proposed method can solve decision making problems with PLTSs, the former failed to handle group decision making problems while the latter can manage both single decision making problems and the group ones. Thereby, the latter has wider application fields.

(2) Practical analysis

The distinguishing power of the proposed method is stronger than that of method (Xu et al.2019), which can be verified by Section 6.2. This is largely due to the fact that the preference functions of the improved PL-PROMETHEE approach proposed in this paper are in the form of PLTSs and can distinguish alternatives neatly, whereas the preference functions defined in method (Xu et al.2019) and the other PL-PROMETHEE approach (Liu and Li, 2018) are all crisp numbers.

(2) The stableness of the proposed method is better than method (Liu and Li, 2019). The former determines DMs’ weights and attribute weights by building objective programming model. Thereby, the decision result is unique under the given decision information. In contrast, the latter assigned DMs’ weights or attribute weights in advance. Hence, the change of such weights may result in different decision results.

(3) Method Gou et al. (2017) is suitable for the environment of hesitant fuzzy linguistic variables. Nevertheless, it cannot handle MAGDM problems with PLTSs which can be solved by the proposed method.

7Conclusions

In today’s internet age, car sharing is more and more popular. The car sharing platform selection is important for tourists, which can be regarded as a MAGDM problem. The PLTS is a powerful tool to represent the evaluation information of DMs in complex MAGDM problems. This paper introduces PLWAM and PLWGM operators firstly and studies some desirable properties of them. Subsequently, a hesitancy index of a PLTS and a general distance measure between PLTSs are defined. Then, a new approach is proposed to rank PLTSs. To measure the fuzziness and hesitancy of a PLTS, a fuzzy entropy and a hesitancy entropy of PLTSs are presented. Afterwards, a total entropy of PLTSs is defined to measure the uncertainty of a PLTS. Meanwhile, a cross entropy between PLTSs is presented. Based on the total entropy and the cross entropy of PLTSs, DMs’ weights and attribute weights are determined objectively. Finally, an improved PL-PROMETHEE method is developed by defining new preference functions and a total preference index. A car sharing platform selection is operated at length to illustrate the applications and advantages of the proposed method.

Apart from solving the selection of car sharing platform, the proposed method can be applied into many decision making fields, such as financial management (Kou 2019a, 2019b) and supplier selection. This paper ignores the risk attitudes of DMs which may play an important role in some finance decision problems. Future study will investigate MAGDM problems with PLTSs considering DMs’ risk attitudes.

Appendices

A

AAppendix

Theorem 1 can be proved by mathematical induction on n as follows:

For n=1, Theorem 1 obviously holds based on (ii) in Definition 9.

Suppose Theorem 1 holds for n=q, which means that:

(A.1)
PLWAX(L1(p),L2(p),,Lq(p))=g1(kj=1,2,,#Lj(p)j=1,2,,q{(1j=1n(1g(Lj(kj)))ωj)(p1(k1)p2(k2)pn(kq))}).

Furthermore, one gets

(A.2)
kj=1,2,#Lj(p)j=1,2,np1(k1)p2(k2)pn(kn)=1.
When n=q+1, one has
(A.3)
PLWAX(L1(p),L2(p),,Lq(p),Lq+1(p))=(ω1L1(p)ω2L2(p)ωqLq(p))ωq+1Lq+1(p).

According to Eq. (A.1) and the operational laws (i.e. (i) and (ii)) in Definition 9, it generates

PLWAX(L1(p),L2(p),,Lq(p),Lq+1(p))=g1(kj=1,2,,#Lj(p)j=1,2,,q{(1j=1n(1g(Lj(kj)))ωj)(p1(k1)p2(k2)pq(kq))})+g1(kq+1=1,2,,#Lq+1(p){(1(1g(Lj(kj))ωq+1)(pq+1(kq+1))})=g1(kj=1,2,,#Lj(p)j=1,2,,qkq+1=1,2,,#Lq+1(p){((1j=1q(1g(Lj(kj)))ωj)+(1(1g(Lq+1(kq+1))ωq+1)(1j=1q(1g(Lj(kj)))ωj×(1(1g(Lq+1(kq+1))ωq+1)))(p1(k1)p2(k2)pq(kq)pq+1(kq+1))}=g1(kj=1,2,,#Lj(p)j=1,2,,q+1{(1j=1q+1(1g(Lj(kj)))ωj)(p1(k1)p2(k2)pq+1(kq+1))}).
Thus, Eq. (4) holds for n=q+1. Therefore, Eq. (5) holds for all n.

On the other hand, consider

kj=1,2,#Lj(p)j=1,2,q,q+1p1(k1)p2(k2)pq(kq)pq+1(kq)=kj=1,2,#Lj(p)j=1,2,qp1(k1)p2(k2)pq(kq)(pq+1(1)+pq+1(2)++pq+1(#Lq+1(p)).

As pq+1(1)+pq+1(2)++pq+1(#Lq+1(p)=1, according to Eq. (A.2), one has

kj=1,2,#Lj(p)j=1,2,q,q+1p1(k1)p2(k2)pq(kq)pq+1(kq)=kj=1,2,#Lj(p)j=1,2,qp1(k1)p2(k2)pq(kq)=1.

Thus, Eq. (5) holds for n=q+1. Hence, Eq. (6) holds for all n.

B

BAppendix

In order to prove that EF(L¯(p)) is a fuzzy entropy, it is necessary only to prove that EF(L¯(p)) satisfies the properties in Definition 15.

By the conditions (ii) and (iii) in Theorem 3, the properties (i), (ii) and (iii) obviously hold.

Next, we only prove the property (iv).

As (L¯(p))(c)=g1(k=1,2,,#L(p){(1g(L¯(k)))(p(k))}), it can be easily deduced that e((L¯(p))(c))=1e((L¯(p))) Therefore, one obtains

(B.1)
EF((L¯(p))(c))=f(e((L¯(p))(c)))=f(1e(L¯(p))).

Since 0e(L¯(p)1 and f(1t)=f(t), one has f(1e(L¯(p)))=f(e(L¯(p))). Hence, in virtue of Eq. (B.1), one gets EF((L¯(p))(c))=EF(L¯(p)).

The proof is completed.

C

CAppendix

In order to prove that h(L¯(p)) is a hesitancy entropy of L¯(p), it is necessary to prove that h(L¯(p)) satisfies three properties in Definition 16.

(i) The sufficiency obviously holds.

Next, we prove the necessity.

When h(L¯(p))=0, suppose #L¯(p)>1. In virtue of Eq. (9), we have

(C.1)
|g(L(k))g¯|=0for anyk=1,2,,#L¯(p).

Therefore, one has g(L(k))=g¯ (k=1,2,,#L¯(p)) and then #L¯(p)=1, which is contrary with the condition #L¯(p)>1. Thus, we obtain #L¯(p)=1. Therefore, it concludes that L¯(p)={sα(1)}.

(ii) When L¯(p)={sτ(0.5),sτ(0.5)}, from Eq. (9), it is clear that h(L¯(p))=1. Furthermore, when h(L¯(p))=1, we derive that 2k=1#L¯(p)|g(L(k))g¯|pi=1. As 0p(k),|g(L(k))g¯|1. Hence, we get p(k)=p(δ)=0.5, L(k)=sτ and L(δ)=sτ. Therefore, L¯(p)={sτ(0.5),sτ(0.5)}.

(iii) Suppose L¯(p)={L(1)(p),L(2)(1p)}, then one obtains g¯=(g(L(1))+g(L(2)))/2. For convenience, let g(L(1))>g(L(2)), we have 2k=1#L¯(p)|g(L(k))g¯|pi=|g(L(1))g(L(2))|. When L(1)L(2), one gets |g(L(1))g(L(2))|0. Therefore, it generates EH(L¯1(p))0.

(iv) This is obviously true.

(v) Since (L(p))(c)=g1(k=1,2,,#L(p){(1g(L(k)))(p(k))}), it is deduced that

EH((L¯(p))(c))=2k=1#L¯(p)(1g(L(k))(1g¯))pk=2k=1#L¯(p)(g(L(k))g¯)pk=EH(L¯(p)).

The proof is completed.

D

DAppendix

In virtue of Eq. (24), one has

CE(L¯1(p),L¯2(p))=k=1#L¯1(p)(A1(k)ln2A1(k)A1(k)+A2(k)(1A1(k))ln2(1A1(k))1A1(k)+1A2(k))=k=1#L¯1(p)(A1(k)lnA1(k)+A2(k)2A1(k)+(1A1(k))ln1A1(k)+1A2(k)2(1A1(k)))

As f(x)=lnx is a concave function, by employing the Jessen inquality, we get

CE(L¯1(p),L¯2(p))k=1#L¯1(p)ln(A1(k)A1(k)+A2(k)2A1(k)+(1A1(k))1A1(k)+1A2(k)2(1A1(k)))=ln1=0.

That is CE(L¯1(p),L¯2(p))0. Therefore, CE(L¯1(p),L¯2(p))0. The equality holds if and only if A1(k)=A2(k)(k=1,2,,#L¯1(p)). Namely, g(L¯1(k))p¯1(k)=g(L¯2(k))p¯2(k) (k=1,2,,#L¯1(p)). Hence, one has e(L¯1(p))=e(L¯2(p)).

The proof is completed.

Acknowledgements

The authors would like to thank the editor in chief and anonymous reviewers for their insightful and constructive comments and suggestions that have led to an improved version of this paper.

References

1 

Bai, C.Z., Zhang, R., Qian, L.X., Wu, Y.N. ((2017) ). Comparisons of probabilistic linguistic term sets for multi-criteria decision making. Knowledge-Based Systems, 119: , 284–291.

2 

Dong, Y.C., Zha, Q.B., Zhang, H.J., Kou, G., Fujita, H., Chiclana, F., Herrera-Viedma, E. ((2018) ). Consensus reaching in social network group decision making: research paradigms and challenges. Knowledge-Based Systems, 162: , 3–13.

3 

Feng, X.Q., Zhang, Q., Jin, L.S. ((2020) ). Aggregation of pragmatic operators to support probabilistic linguistic multi-criteria group decision-making problems. Soft Computing, 24: (10), 7735–7755.

4 

Gou, X.J., Xu, Z.S. ((2016) ). Novel basic operational laws for linguistic terms, hesitant fuzzy linguistic term sets and probabilistic linguistic term sets. Information Sciences, 372: , 407–427.

5 

Gou, X.J., Xu, Z.S., Liao, H.C. ((2017) ). Hesitant fuzzy linguistic entropy and cross-entropy measures and alternative queuing method for multiple criteria decision making. Information Sciences, 388–389: , 225–246.

6 

Halouani, N., Chabchoub, H., Martel, J.M. ((2009) ). PROMETHEE-MD-2T method for project selection. European Journal of Operational Research, 195: (3), 841–849.

7 

Jiang, L.S., Liao, H.C. ((2020) ). Mixed fuzzy least absolute regression analysis with qualitative and probabilistic linguistic information. Fuzzy Sets and Systems, 387: , 35–48.

8 

Klement, E.P., Mesiar, R., Pap, E. ((2000) ). Triangular Norms. Kluwer Academic Publishers, Dordrecht.

9 

Krishankumar, R., Ravichandran, K.S., Saeid, A.B. ((2017) ). A new extension to PROMETHEE under intuitionistic fuzzy environment for solving supplier selection problem with linguistic preferences. Applied Soft Computing, 60: , 564–576.

10 

Kou, G. (2019a). Introduction to the special issue on FinTech. Financial Innovation, 5(45). https://doi.org/10.1186/s40854-019-0161-1.

11 

Kou, G. (2019b). Editor’s introduction. Financial Innovation, 5(46). https://doi.org/10.1186/s40854-019-0160-2.

12 

Kou, G., Yang, P., Peng, Y., Xiao, F., Chen, Y., Alsaadi, F.E. (2020). Evaluation of feature selection methods for text classification with small datasets using multiple criteria decision-making methods. Applied Soft Computing, 86. https://doi.org/10.1016/j.asoc.2019.105836.

13 

Li, P., Wei, C.P. ((2019) ). An emergency decision-making method based on D-S evidence theory for probabilistic linguistic term sets. International Journal of Disaster Risk Reduction, 37: , 101178. https://doi.org/10.1016/j.ijdrr.2019.101178.

14 

Li, Y., Zhang, Y.X., Xu, Z.S. ((2020) ). A decision-making model under probabilistic linguistic circumstances with unknown criteria weights for online customer reviews. International Journal of Fuzzy Systems, 22: (3), 777–789.

15 

Liang, D.C., Kobina, A., Quan, W. ((2018) ). Grey relational analysis method for probabilistic linguistic multi-criteria group decision-making based on geometric Bonferroni Mean. International Journal of fuzzy Systems, 20: (7), 2234–2244.

16 

Liao, H.C., Jiang, L.S., Lev, B., Fujita, H. ((2019) ). Novel operations of PLTSs based on the disparity degrees of linguistic terms and their use in designing the probabilistic linguistic ELECTRE III method. Applied Soft Computing, 80: , 450–464.

17 

Liao, H.C., Mi, X.M., Xu, Z.S. ((2020) ). A survey of decision-making methods with probabilistic linguistic information: bibliometrics, preliminaries, methodologies, applications and future directions. Fuzzy Optimization and Decision Making, 19: (1), 81–134.

18 

Lin, M.W., Chen, Z.Y., Liao, H.C., Xu, Z.S. ((2019) ). ELECTRE II method to deal with probabilistic linguistic term sets and its application to edge computing. Nonlinear Dynamics, 96: (3), 2125–2143.

19 

Liu, P.D., Li, Y. ((2018) b). The PROMETHEE II method based on probabilistic linguistic information and their application to decision making. Informatica, 29: (2), 303–320.

20 

Liu, P.D., Li, Y. ((2019) ). An extended MULTIMOORA method for probabilistic linguistic multi-criteria group decision making based on prospect theory. Computers & Industrial Engineering, 136: , 528–545.

21 

Liu, P.D., Teng, F. ((2019) ). Probabilistic linguistic TODIM method for selecting products through online product reviews. Information Sciences, 485: , 441–455.

22 

Liu, H.B., Jiang, L., Xu, Z.S. ((2018) ). Entropy measures of probabilistic linguistic term sets. International Journal of Computational Intelligence Systems, 11: (1), 45–57.

23 

Liu, P.D., Li, Y., Teng, F. ((2019) ). Bidirectional projection method for probabilistic linguistic multi-criteria group decision-making based on power average operator. International Journal of Fuzzy Systems, 21: (8), 2340–2353.

24 

Mao, X.B., Wu, M., Dong, J.Y., Wan, S.P., Jin, Z. ((2019) ). A new method for probabilistic linguistic multi-attribute group decision making: application to the selection of financial technologies. Applied Soft Computing, 77: , 155–175.

25 

Mi, X.M., Liao, H.C., Wu, X.L., Xu, Z.S. ((2020) ). Probabilistic linguistic information fusion: a survey on aggregation operators in terms of principles, definitions, classifications, applications, and challenges. International Journal of Intelligent Systems, 35: (3), 529–556.

26 

Pang, Q., Xu, Z.S., Wang, H. ((2016) ). Probabilistic linguistic term sets in multi-attribute group decision making. Information Sciences, 369: , 128–143.

27 

Peng, H.G., Wang, J.Q., Zhang, H.Y. (2020). Multi-criteria outranking method based on probability distribution with probabilistic linguistic information. Computers & Industrial Engineering, 141. https://doi.org/10.1016/j.cie.2020.106318.

28 

Rodríguez, R.M., Martínez, L., Herrera, F. ((2012) ). Hesitant fuzzy linguistic term sets for decision making. IEEE Transaction on Fuzzy Systerms, 20: (1), 109–119.

29 

Wan, S.P., Xu, G.L., Wang, F., Dong, J.Y. ((2015) ). A new method for Atanassov’s interval-valued intuitionistic fuzzy MAGDM with incomplete attribute weight information. Information Sciences, 316: , 329–347.

30 

Wu, X.L., Liao, H.C. ((2018) ). An approach to quality function deployment based on probabilistic linguistic term sets and ORESTE method for multi-expert multi-criteria decision making. Information Fusion, 43: , 13–26.

31 

Wu, X.L., Liao, H.C. ((2019) ). A consensus-based probabilistic linguistic gained and lost dominance score method. European Journal of Operational Research, 272: (3), 1017–1027.

32 

Wu, X.L., Liao, H.C., Xu, Z.S., Hafezalkotob, A., Herrera, F. ((2018) ). Probabilistic linguistic MULTIMOORA: a multi-criteria decision making method based on the probabilistic linguistic expectation function and the improved Borda rule. IEEE transactions on Fuzzy Systems, 26: , 3688–3702.

33 

Wu, X.L., Zhang, C., Jiang, L.S., Chang, H.L. ((2019) ). An integrated method with PROMETHEE and conflict analysis for qualitative and quantitative decision-making: case study of site selection for wind power plants. Cognitive Computation, 272: (1), 1017–1027.

34 

Xu, Z.S. ((2004) ). A method based on linguistic aggregation operators for group decision making with linguistic preference relations. Information Sciences, 166: (1), 19–30.

35 

Xu, Z.S. ((2005) ). Deviation measures of linguistic preference relations in group decision making. Omega, 33: (3), 249–254.

36 

Xu, G.L., Wan, S.P., Wang, F., Dong, J.Y., Zeng, Y.F. ((2016) ). Mathematical programming methods for consistency and consensus in group decision making with intuitionistic fuzzy preference relations. Knowledge-Based Systems, 98: , 30–43.

37 

Xu, Z.S., Luo, S.Q., Liao, H.C. ((2019) ). A probabilistic linguistic PROMETHEE method and its application in medical service. Journal of System and Engineering, 34: , 760–769.

38 

Zadeh, L.A. ((1975) ). The concept of a linguistic variable and its application to approximate reasoning-II. Information Sciences, 8: , 301–353.

39 

Zhang, X.L. ((2018) ). A novel probabilistic linguistic approach for large-scale group decision making with incomplete weight information. International Journal of Fuzzy Systems, 20: (7), 2245–2256.

40 

Zhang, Y.X., Xu, Z.S., Wang, H., Liao, H. ((2016) ). Consistency-based risk assessment with probabilistic linguistic preference relation. Applied Soft Computing, 49: , 817–833.

41 

Zhang, Y.X., Xu, Z.S., Liao, H.C. ((2017) ). A consensus process for group decision making with probabilistic linguistic preference relations. Information Sciences, 414: , 260–275.

42 

Zhang, H.H., Kou, G., Peng, Y. ((2019) ). Soft consensus cost models for group decision making and economic interpretations. European Journal of Operational Research, 277: (3), 964–980.

43 

Zhang, X.F., Gou, X.J., Xu, Z.S., Liao, H.C. ((2019) ). A projection method for multiple attribute group decision making with probabilistic linguistic term sets. International Journal of Machine Learning and Cybernetics, 10: (9), 2515–2528.