Abstract
1. Introduction
Vehicular Ad hoc networks (VANETs) are one important type of the mobile ad hoc networks (MANETs) developed as the basis of Intelligent Transportation Systems (ITS) to provide safer, better, and more efficient road conditions. In VANETs, the main network nodes are the smart vehicles and the road-side infrastructure units (RSUs) that are able to communicate with each other through vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I) communications. Such communications provide a variety of applications ranging from exchanging life-saving information, such as environmental and driving hazards, to traffic congestion, touristic messages, and advertisements.
The V2V and V2I communications are at short distance and high speed in VANETs. Because the communication range of vehicles is limited, packets in VANETs need other vehicles to forward cooperatively. However, in real application scenarios, the vehicles are driven by humans and the human behavioral tendencies are reflected in the behaviors of the nodes. In the event of high-energy consumption and low bandwidth availability, some of the vehicle nodes in the network might refuse to forward other's message packets. Such nodes are called as selfish nodes and they always intend to maximize their own profit, causing undesirable delays in the message delivery and increase the network latency, which in turn affects the entire performance of the network. If a large number of selfish vehicles exist in VANETs, the performance and function of VANETs will be greatly influenced. For these reasons, it is essential to detect selfishness and encourage them in order to promote performance of VANETs. Reputation mechanism is pivotal for node cooperation in packet forwarding in the large-scale vehicular ad hoc networks [1–3]. Through reputation evaluation, selfish behaviors can be discovered and punished in a certain degree.
In order to monitor and isolate misbehaving of packet forwarding in VANETs, a number of reputation technologies have developed to manage reputations and limit the negative impact of selfish nature. For an observed vehicle node
The network traffic consumption by continually reporting reputation data is another major concern problem. In VANETs, vehicle nodes belong to different individual. The users are not willing to consume extra traffic or resources for periodically reporting local reputations in order to avoid influencing personal applications. In addition, reputations from different recommenders on the different period may cause local reputations confusion (see Section 3.2). Calculating reputation based on a reasonable reporting cycle is essential for saving network traffic and resources and making reputation more accurate.
Reputation management needs to better suit VANETs and increase the reputation accuracy, and thus we propose a hierarchical reputation evidence decision system (REDS) based on the Dempster-Shafer evidence theory. The main merit of Dempster-Shafer evidence method [6, 7] is that it combines different evidences, especially some uncertainties from different evidence sources. Uncertain evidences are used to denote reputations caused by vehicle mobility and channel noise and further promote reputation accuracy and mitigate the adverse impact of network characteristics.
REDS is a dynamic three-layer reputation evidence combination structure. In the lowest coordination observer decision layer, some of
One of the advantages of the paper is that colluders can also be detected through reputation evidence feedback. Colluders usually report higher reputations for each other and decrease other's reputations. To avoid collusion, the host manager feeds the ultimate reputation back to different layers' evaluators. The upper evaluator can manage the trustworthy degree of evaluators of lower layer. The trustworthy degrees of evidence directly influence evaluators' credits. If an evaluator's reputation evidences are always deviated from others, it tends to be a colluder, and then its credit is also negative in the end.
In addition, the responsibility for transmitting reputation evidences leads evaluators to consume more network traffic. So we design an adaptive evidence gathering cycle in views of vehicles' mobility and personal applications.
In detail, our contributions are as follows.
The dynamic three-layer reputation evidence combination system: in the current reputation management technologies [8–10], the reputation manager ignores the reputation which deviates dramatically from the average value in order to avoid falsified reputation. It leads the accurate reputation to be abandoned when large-scale collusions exist. Our proposed scheme combines Dempster-Shafer evidence with hierarchical evidence clustering on the dissimilar evaluation layer to calculate more accurate reputations and detect selfish vehicle nodes in less time than other typical reputation algorithms. Especially it calculates the trust degree of evaluators and therefore effectively isolates collusive coordination observers that conspiratorially report higher reputation for selfish nodes. Adaptive reputation gathering cycle: too frequent polling reputation evidences will cost plenty of network traffic. An adaptive reputation gathering cycle based on Weber-Fechner's law is proposed in order to maintain the stability of reputation management system and save network traffic. Considering vehicles' movement and personal applications, dynamic changing cycle can seek a balance between reputation valuation accuracy and network traffic.
The rest of papers are organized as follows. Section 2 provides some related works about the reputation cooperation. Section 3 introduces our hierarchical reputation evidence decision system (REDS), and Section 4 depicts the simulation results with respect to typical reputation evaluation algorithms and demonstrates our algorithm's efficiency and superiority. In Section 5, we conclude the paper.
2. Related Works
Some research works have been done on vehicle stimulation cooperation in VANETs at present. Related research of MANETs is relatively abundant. The stimulation cooperation mechanisms in the MANETs are generally classified into virtual currency-based and reputation-based stimulation system. In the virtual currency system, nodes pay virtual money (called nuglets) [11] for forwarding service. In detail, a source pays for each forwarding node on the routing path. And each node also needs to provide forwarding service to gain nuglets. The previously mentioned price mechanism is not flexible enough to maximize cooperators' profits. For example, a source hardly evaluates accurately the total of nuglets in the dynamic changing network environment. Some researchers introduce economic incentive technology in the self-organized ad hoc networks. To maximize the payoff of relay nodes, Ji et al. [12] propose a game pricing-based routing mechanism. Packet-forwarding services are negotiated and auctioned in the orthogonal frequency-division multiplexing setting. Mukherjee and Kwon [13] design a robust multiobject bundled auction method for simultaneous partner selection. After successful auction, a contractor (winning bidder) achieves satisfied QoS and rewards the service provider (seller).
Reputation management systems are adopted to detect and isolate misbehaving nodes to release their negative influence on the network performance. In the light of adverse impact of misbehaving nodes, the paper [14] presents the watchdog and detects uncooperative relay nodes. Detecting node maintains a buffer to store the number of neighbors' received and sent packets. But it is probably influenced by channel collision. MobiGame [15] designs a user-centric reputation incentive system for delay tolerant network (DTN). Packet forwarding's cost and reward can get a Bayesian equilibrium through a game-theoretic scheme. Refaei et al. [1] represent a time-slotted evaluation approach to detect timely node's changing behaviors. In addition, the sequential probability ratio test (SRRT) is used for judging whether a neighbor is selfish. Anantvalee and Wu [9] define two thresholds,
For the reputation store, DHT Trust overlay network [16] uses CHORD to distribute local reputation to reputation manager. Finger nodes issue reputation feedback in the certain interval and fast aggregate local reputation into the global reputation. Li and Shen [8] propose the account-aided reputation management system (ARM) to stimulate selfish mobile nodes. The ARM system calculates reputations and credits to distinguish selfishness from cooperation. In particular, the authors introduce the distributed hash table (DHT) to store circulating reputations and credits. As node mobility, DHT uses a lightweight maintenance protocol to reduce the number of reputation structure reestablishment.
Although the above-mentioned references have provided incentive mechanisms for selfish nodes in the wireless self-organized network, there are some deficiencies that need to be solved further, for example, reputation imprecision due to vehicles' mobility, reputation falsifying caused by collusion, and so forth. And thus, our REDS effectively promotes reputation accuracy by using hierarchical reputation evidence decision and distinguishes suspect collusion vehicle nodes and decreases their trustworthy degree through reputation feedback.
3. Reputation Evidence Decision System
In the paper, we propose a dynamic three-layer reputation evidence decision system. As illustrated in Figures 1(a) and 1(b), the hierarchical structure contains all of mobile vehicles that are classified into 4 roles: host manager (HM), coordination manager (CM), coordination observer (CO), and observed nodes. An observed node

(a) Reputation-based network topology. (b) Three-layer reputation decision architecture.
In addition,
However, if a CO or a CM runs out of the observed node's communication range, it downgrades itself to a normal node, and will be no longer engaged in reputation gathering in the current round until it enters the observed node's communication range again. In a reputation gathering cycle, each CO chooses the nearest CM in
An example depicted in Figures 1(a) and 1(b) is given as follows. It is assumed that the observed vehicle node
In the middle-layer of reputation decision,
Some reputation evaluation algorithms ignored the high deviated reputation which may be caused by misreport or falsification. On the one hand, the benign reputation may be abandoned in the large-scale collusion environment and then the benign node is classified into selfish. On the other hand, they have no restraint measures to punish colluders with falsified reports in many times. These falsified or misreporting nodes are unsuitable to continue to serve as COs. So we introduce the Dempster-Shafer Evidence theory [17, 18] into reputation decision in order to calculate reputation more accurately. Through reputation feedback on COs and CMs, we can fast detect selfish node and colluders.
Moreover, we design a dynamic reporting cycle for coordination managers and coordination observers according to their mobility and resources. This is because they will consume extra network traffic for collecting and merging evidences. However mobile vehicles belong to different individual users, and then the extra network traffic consumption will influence user's personal applications. It means the reputation managers should save extra network traffic consumption.
3.1. Reputation Evidence Combination
In the Dempster-Shafer evidence reasoning mechanism, evidences are denoted as some possible events. Using reasoning combination rule to aggregate multiple belief evidences under uncertainty. To calculate an accurate comprehensive reputation, we consider effect of each reputation evidence rather than simply ignore the value that deviates largely from the average. We design the dynamic three-layer reputation evidence aggregation structure for reputation evaluator at different layers. And furthermore, the system executes reputation feedback to amend each evaluator's trust weight and distinguish colluders.
In the coordination observer reputation decision layer, a set of hypotheses about observed forwarding behaviors is denoted as a frame of discernment
Observer
The BPA for proposition
If less, it means that
Here the discount factor
A coordination observer may be colluder that reports a higher reputation for selfish nodes. Once a CO is detected to be a colluder, it will be not suitable for reputation gathering and decision even though it still runs within the observed node's communication range. These collusive COs will be punished and isolated and will not be permitted to upgrade themselves to be CMs. To detect colluders,

Feedback for trustworthy weights.
Define the direction cosine between any evidence by any evaluators
The trust weights of evaluators (coordination observers and managers) are defined as follows:
Through evaluation of trust weights of COs and CMs, a reputation evidence reported by a CO or a CM will be influenced:
Define the reputation evidence combination rule based on reputation evidence feedback as follows:
Here
The combined evidence represents the probability of event
To further avoid evidence conflict, we improve D-S evidence rule integrated with hierarchical clustering theory besides adding trust weight for each reputation evaluator. In the coordination manager reputation decision layer, a coordination manager receives the coordination observers' belief reputation value in its own domain. If some evidence equals 0, then we use the hierarchical clustering mechanism to combine evidences to avoid evidence conflict, or else it uses weighted D-S evidence combination rule.
It is assumed that the coordination manager acquires
Assume that another evidence cluster
The pseudocode of reputation evidence calculation is shown in Pseudocode 1.
Calculate Combine reputation evidences based on hierarchical clustering (formula (10)–(12)); Use reputation regarded as a new gravity center if using hierarchical evidence clustering next combination.
Pseudocode 1
In the host manager reputation decision layer, the host manager needs to know the relative degree about each coordination manager when evidence combination. Namely, it calculates the direction cosine between
High-reputed and low-reputed nodes should be rewarded or punished. Define a selfish character factor
Here
Colluders report high reputation for selfish nodes and low reputation for cooperators, so it also expects its ally to report high reputation in return. It can be regarded as selfish in certain extent. Given a trust weight of the coordination observer
Here
Here
3.2. Adaptive Reputation Gathering Cycle
However polling may cause more network traffic consumption if vehicle nodes gather reputation more frequently. In VANETs, selfish vehicles [19, 20] pursue the least cost to reduce network traffic consumption. But the responsibility for gathering and merging reputation leads evaluator to consume more. It should seek a balance between gathering frequency and consumption of network traffic. Therefore, each neighbor vehicle node which runs within the range of
In the paper, we improve Weber-Fechner's law [21] in order to update gathering cycle that is subject to a variety of environment stimulating factors, such as the number of current applications and mobility of vehicles. Weber-Fechner's law uses a linear function of logarithm for describing the relationship between individual's response and incitement due to external environments. In the following formula,
Assume that
4. Simulation Analysis
We conduct simulations to demonstrate the performance of REDS in our paper. The simulated VANETs include 60/120 vehicle nodes randomly deployed in the 1 km driveway area. The velocity of each vehicle is dynamically changed and is set randomly from 50 km/h to 100 km/h. As shown in Figure 1, the RSUs are deployed along the road and regarded to be HMs. Different roles of vehicle nodes such as COs or CMs are determined by vehicle nodes themselves according to the dynamic changing gathering cycles and vehicle's velocities. We assume all of managers are trustworthy, while coordination observers maybe collude with observed nodes in order to report high reputation evidences for them. Colluders are divided into random collusion and group collusion. Group collusion means all of coordination observers in a group advisedly overwhelm other groups. The initial credits of all of observed nodes (ONs) and COs are equal to 2. Physical bandwidth is set to 2 Mbit/s. We compare our algorithm with ARM [8] and purely D-S Proof Fusion mechanism.
The ARM system calculates reputations and credits to distinguish selfishness from cooperation. In particular, the authors introduce the distributed hash table (DHT) to store circulating reputations and credits. As node mobility, DHT uses a lightweight maintenance protocol to reduce the number of reputation structure reestablishment.
Firstly we verify the validity of selfish node detection when the network size changes. In Figure 3, there are 10 selfish observed nodes (ONs) and 10 cooperative ONs. In the communication ranges of these observed nodes, we set 15 collusive nodes and 25 cooperative nodes (include ONs' host managers). We assume that all collusive nodes and cooperative nodes have qualification to be COs, whereas all managers (including HMs and CMs) need to be cooperative and trustworthy. That is to say, a collusive CO cannot upgrade itself to a CM even though it runs within the corresponding observed node's communication range all the time.

The time for detecting all selfish observed nodes.
Figure 3 shows the time for detecting all of selfish observed nodes. Here selfish degrees (ON's drop packet rate) are defined in [0.6, 0.9]. REDS has the minimum detection time for all of selfish nodes. The greater the selfish degree is, the less the detection time is taken. While in the figure we see a comparative result in D-S when no one ticket veto situation happens, and ARM shows the worst performance in this situation. As ARM only simply ignores anomalous reputations rather than punishing colluders, it is difficult to distinguish cooperators from collusive COs in the large-scale group collusion. And then ARM consumes the longest time for detecting selfishness.
In Figure 4, we define, respectively, the selfish degree of the coordination observer nodes in [0.8, 1] (Figure 4(a)) and [0, 0.2] (Figure 4(b)) and evaluate their credits according to the simulation setting as Figure 3. Each ON's initial credit is set to 2. For selfish ONs in REDS, the average credit decreases slowly in the start detection stage. That is because the reputation feedback has not distinguished all of collusive COs. As time goes on, collusive COs are detected one by one (shown in Figures 6–9). Based on this, the selfish ON's credits (shown in Figure 4(a)) decrease quickly because less collusive COs report fake high reputation evidences. REDS and DS act more steeply which means that they detect selfishness more quickly than ARM. ARM puts selfish nodes into blacklist but has no ability to detect collusive COs. The selfish nodes' reputation evaluations are always influenced by falsified high reputation reported by collusive COs.

(a) Average credit of selfish ONs. (b) Average credit of cooperative ONs.
Figure 4(b) represents average credit of cooperative ONs. Selfish observed nodes are detected as quickly as possible in the REDS, and thus the system has a higher throughput than others. As more trustworthy nodes are selected to be forwarding node, the credits of cooperative ONs have stably increased compared with DS and ARM.
In Figure 5, we evaluate whether the average credit of REDS would be influenced by the network size or not. The simulation executes 33 minutes, respectively, with different network size (60 or 120). The number of selfish nodes is 1/6 of the total. The total of cooperative ONs follows the number. We can see that the average credit of selfish observed nodes has decreased to zero. The credit's slope with 60 nodes is similar with 120 nodes, which means REDS executes stably regardless of the network size. When the rate of number of COs to number of ONs is changeless, the declining ratio of credit is also stationary.

Average credits change over different network size.

The detection time for collusive COs.

Average credit of group colluders.

Average credit of colluders over different network size.

Network traffic consumption.
We will verify the effectiveness of collusive COs' detection in REDS when the network size is 60. Define 15 collusive nodes are randomly distributed and they have equal change to be COs. In each coordination managers' domain, 2~3 random colluders are included at most in the random collusion mode. Group collusion mode represents the network including several random colluders and 1~2 collusion groups. All of coordination observers in each collusion group agree with selfish observed nodes to report falsified high reputation evidences.
Figure 6 demonstrates the detection time for collusive coordination observers. We can see that group collusion detection is slightly faster than random collusion in REDS. Through reputation feedbacks on COs and CMs, REDS amends their trust weights. We have assumed that all of coordination managers are trustworthy. As to group collusion, it is easier for CM to evaluate the behaviors of all of coordination observers in its own management domain.
Figures 7 and 8 verify the stability of REDS in the conditions of random collusion and group collusion. The network size is 60, and the total of colluders (including group and random colluders) is 15. Figure 7 represents the average credit of colluders when the number of group colluders is 10 (5 random collusion, in addition) or 15 (no random collusion in the network). We can see that average credit decreasing speed of 15 group colluders is 0.088 per minute, slightly faster than 10 group colluders (0.069 per minute). According to the results of Figure 6, group colluders are detected in less time due to the trustworthy weights of group colluders and their high-level coordination managers are always lower than others.
Figure 8 presents observation results of average credit of colluders over different network size (60 or 120). If all collusive vehicles become COs, then the number of collusive COs is 1/4 of the total. We adopt reputation feedback and credit update in REDS, and each observed node insures more than 5 COs to detect itself though the network size enlarges. REDS detects colluders in a stable speed and approximately regardless of network size in the distributed conditions.
Figure 9 represents the network traffic consumption in different reputation evaluation system. The simulation settings are followed as in Figure 3. The packets size of reputation evidence and weight feedback is 512 bytes. For simplicity, the following algorithms do not add more nodes for new ONs after removed selfish ONs and collusive COs to blacklist. It means that they only continue to detect old cooperative ONs.
Curves of DS and ARM have the same gradient and
5. Conclusion
In the paper, we propose a three-layer reputation evidence decision system (REDS) to detect misbehaving nodes in VANETs. REDS can distinguish fraudulent information from real reputation evidences and avoid credits of cooperative nodes being affected by falsified information. Collusive coordination observers usually conspiratorially report fraudulent reputation evidences in a random or group collusion way. If only ignores highly deviated information rather than punishes premeditated reporters, collusion will always exist. We feed back trust degree of each coordination observer to its coordination manager, thus helping reputation evidence combination and collusion detection. The credits of the coordination observers decreased or increased according to the results of their trust degree weights. Moreover, an adaptive reputation evidence gathering cycle is proposed to replace frequent polling mechanism and save the network traffic. The simulation results demonstrate REDS having high performance of detection for selfish and collusive behaviors.
