This paper studies the distributed energy efficient resource management in the Internet of Things (IoT). Wireless communication networks support the IoT without limitation of distance and location, which significantly impels its development. We study the communication channel and energy management in the wireless communication network supported IoT to improve the ability of connection, communication, share and collaboration, by using the game theory and distributed learning algorithm. First, we formulate an energy efficient neighbor collaborative game model and prove that the proposed game is an exact potential game. Second, we design a distributed energy efficient channel selection learning algorithm to obtain the global optimum in a distributed manner. We prove that the proposed algorithm will asymptotically converge to the global optimum with geometric speed. Finally, we make the simulations to verify the theoretic analysis and the performance of proposed algorithm.
1. Introduction
T
he Internet of Things (IoT) has drawn much attention from academia as well as industry recently. In 1999, Kevin Ashton proposed the term IoT. With the development of internet and wireless communication, the IoT has been seen as an important technological revolution that brings us into a new ubiquitous connectivity, computing, and communication era
[1]
. The possible applications of IoT are significant, such as climate monitoring, transport safety, home automation, health care, supply chain, agriculture, rural development, border security and military application. In IoT, the objects (Obs, including people and things) are expected to have the capabilities of sensing the physical objects, communication, and operating smartly without human intervention. The major technologies that would dominate IoT applications include wireless sensor networks (WSN) and mobile communications
[2]
.
Collaboration is an important feature of IoT. The IoT is widely regarded as global network which allows communication and collaboration between people and things, and between things themselves
[3]
. Wire connected Obs could share information via internet
[4]
, however, much more Obs need wireless connection to overcome limitation of distance and location. The fast wireless communication could support the widely collaboration in the IoT through wireless networks, which significantly impels the development of IoT. Researches on the wireless communication based collaboration in the IoT should be paid much attention to.
In fact, collaboration in IoT needs huge information exchange, which would consume resources, such as communication channel, bandwidth and energy. In practical collaboration via wireless network, the channel resource is very important and limited, then the channel resource management is an important problem. In addition, energy is very sensitive to the IoT
[3]
, especially for the sensors.
There are some researches on the technologies in IoT. Authors in
[5

7]
focued on the surroundings sensing in IoT based on the sensor network technologies. The identification in IoT was studied in
[8

10]
. Authors in
[11

12]
studied the communication protocols. The tracking effects and energy consumption optimization was studied in
[13]
. In all, to the best of our knowledge, there are only limited researches which focused on the energy optimization in IoT, and channel resource management in the IoT has not been paid attention to.
Considering the importance of energy sensitive channel resource management in the collaboration via wireless network, in this paper, we focus on the wireless channel selection for the Obs, towards the energy efficiency. Due to the dense and dynamically deployment of objects in the IoT
[14]
, the centralized control in network management will be highly inefficient for the high complexity.
The game theory
[15]
is a powerful tool to study the interactions among multiple users and has been used in the distributed optimization in many researches about the distributed network
[16

17]
. For some good properties to the distributed global potimization, the potential game was also applied in some researches. In
[18]
, the authors proposed a multicell coordination approach to mitigate the mutual interference among base stations in the frequency slotted cellular networks. Authors in
[19]
proposed local altruistic game and local congestion game which belong to local interaction game in cognitive radio networks. In
[20]
, the authors investigated the problem of joint base station selection and resource allocation in an orthogonal frequency division multiple access heterogeneous cellular network, analyzed this problem by using potential game theoretic approaches, and proposed two different variants of Maxlogit learning algorithms which achieved outstanding performances.
For the IoT, there are only few game based researches to our best knowledge. Authors in
[14]
proposed an energy aware trust derivation scheme using game theoretic approach, which managed overhead while maintaining adequate security of WSNs. In
[3]
, a service providing model was built by using a differential game model. The game solution was gotten in the condition of grand coalition, feedback Nash equilibrium and intermediate coalitions and an allocation policy was obtained by Shapley theory. Nevertheless, with the distributed global energy consumption optimization in IoT which we focus on, related studies are very limited.
To sum up, the multiObs energy consumption optimization is an important and challenging problem in IoT, and the distributed optimization approach is potential but not paid much attention to until now. In this paper, we solve the challenging multiObs energy efficient channel selection problem in IoT in the distributed optimization approach to achieve global optimum, by using the potential game theory and distributed learning algorithm. First, we formulate an energy efficient neighbor collaborative game model and prove that the proposed game is an exact potential game. Second, we design a distributed energy efficient channel selection learning algorithm to obtain the global optimization in a distributed manner. Thirdly, we analyze the converging speed and the computing complexity of the proposed algorithm. We prove that the proposed algorithm will asymptotically converge to the global optimum with geometric speed. Finally, we make the simulations to verify the theoretic analysis and the performance of proposed algorithm.
The rest of this paper is organized as follows: In Section II, we present the system model and problem formulation. In Section III, we formulate the proposed game model, and investigate its properties. In Section IV, we propose the distributed learning algorithm, analysis the convergence of the algorithm, the converging speed and the computing complexity. In Section V, simulation results and discussion are presented. Finally, we provide conclusions in Section VI.
2. System Model and Problem Formulation
We consider a local interactive network
[21]
consisting of
M
objects (Obs), which are connected via various wireless communication technologies and the internet. Obs could transmit their data via 3G mobile communication channels, WiFi channels, Bluetooth and so on. Generally speaking, Obs use wireless frequency channels to communicate. Local interactive is a feature of the IoT that the Obs are always spatially distributed and the effect of each Ob’s action is limited between the neighboring Obs
[1]
. For example, in
Fig. 1
, the data transmission of
Ob_{2}
and
Ob_{3}
would collide if they use the same frequency channel for their neighboring relationship.
Ob_{3}
and
Ob_{6}
could use the same frequency channel without collision because they are not neighbors.
System model
Importantly, because the qualities of wireless channels are related to the location and channel frequency, choosing different frequency channel means different energy consumption. Denote
S_{CH}
=
{1,2,...,N}
as the set of channels, and denote
S_{Ob}
=
{1,2,...,M}
as the set of Obs. If
Ob_{m}
chooses channel
n
, according to the Shannon equation, the obtained capacity would be given by:
where
W_{n}
is the bandwidth of channel
n
,
N_{0}
is the noise power spectrum density,
P_{m}
is the transmitting power of
Ob_{m}
,
d_{m}
is the distance between
Ob_{m}
and its destination of transmitting data.
γ_{n}
is the path loss exponent of channel
n
, and
θ_{n,m}
is the instantaneous random component of the path loss
[22]
on channel
n
for
Ob_{m}
.
Then the transmission power of
Ob_{m}
should be given by:
Due to the lack of central control, collisions may occur when there may be more than one Obs choose the same channel in local region. The typical slotted Aloha transmission mechanism is adopted to reduce the collision, where each Ob accesses the channel with probability
α
. The data transmission of
Ob_{m}
on channel
n
would be successful when no other neighboring Obs attend to access this channel. Denote Ω
_{m}
as the set of
Ob_{m}
’s neighbors (including
Ob_{m}
itself), and Ω
_{m}
 is the number of
Ob_{m}
’s neighbors. The probability of successful transmission is given by:
For each Ob, the higher probability of successful transmission and also the lower transmission power are expected. We define the reward of Ob, e.g.,
Ob_{m}
as follow:
In the perspective of network, the optimization objective is to maximize sum of Ob’s utilities:
Remark:
From the aspect of each Ob, there is a tradeoff for between channel quality and collision probability when making the decision in the network consist of multi Obs. Good channels, i.e., channels which consume lower power, are more likely to be chosen by other Obs, which reduces to higher probability of collision. From the aspect of network, the optimization of network utility and stability are expected. The global optimization of network utility could be obtained in a centralized manner by the exhaustive search. However, the computation complexity is extremely high. For example, even in a relative small scenario where N=5 and M=30, the number of possible Obs’ strategy profiles is 5
^{30}
=9.3132 ×10
^{20}
. The desired method is the one which obtains the global optimization in a distributed manner to low the computation complexity. We solve this problem by using the game theory
[15]
, which is one of the important mathematical tools that analyze interactions among multi objects.
3. The Energy Efficient Neighbor Collaborative Game
In this section, we study the distributed optimization of the formulated energy efficiency problem by using game theory. Every Ob is regarded as a player in the game, and we define the energy efficient neighbor collaborative game (EENCG) as:
where
S_{CH}
is the set of channels,
S_{Ob}
is the set of Obs,
A_{m}
is the available strategy set of
Ob_{m}
∈
S_{Ob}
, where
a_{m}
∈
A_{m}
is the chosen action of
Ob_{m}
.
u_{m}
is the utility function of player
Ob_{m}
, and
u_{m}
(
a_{m}
,
a_{−m}
) denotes the
Ob_{m}
’s utility where
a_{m}
is the action of player
Ob_{m}
and
a_{−m}
is the action profile of other players.
Based on the feature of collaboration in the IoT
[3]
, motivated by local collaboration in biographical systems
[30]
,
[31]
and the collaboration design in networks
[18

20]
,
[23

24]
,
[27]
, we define the following utility function of player, e.g.,
Ob_{m}
, to obtain the global optimization in a distributed decision manner.
where
φ_{i}
is the reward of player
Ob_{i}
and Ω
_{m}
is the set of neighbors of
Ob_{m}
.
Definition 1:
Denote an action profile of players as
a
= {
a
_{1}
,
a
_{2}
,...,
a_{M}
}, and an action profile
is a pure strategy Nash equilibrium (NE) if and only if no player could improve its utility by deviating unilaterally, i.e.,
Theorem 1:
the global optimal solution to the proposed energy efficient channel selection problem (5) is a pure strategy NE of the proposed EENCG game.
Proof:
The following proof is based on the potential game theory
[23]
. We define the network utility as the potential function of the EENCG:
When
Ob_{m}
unilaterally changes its action from
a_{m}
to
and other Obs hold their original strategies, the value of potential function would be changed：
Note that the mutual influence would not occur when the Obs are not neighbors (
Fig. 1
). Then for
Ob_{i}
∈ {
S_{Ob}
╲Ω
_{m}
}, which is not the neighbor of
Ob_{m}
, there will be no influence from the
Ob_{m}
’s action change, i.e.,
Then we have
The result shows that the change in the potential function equals to the change in individual utility function. Thus, according to the theory in
[18]
, the proposed EENCG game is an exact potential game which has at least one pure strategy NE point, and the potential function is Γ(
a_{m}
,
a_{−m}
) =
φ_{net}
.
Furthermore, according to
[23]
, any global or local maxima of the potential function should be a pure strategy NE in exact potential game. Note that, the proposed potential function equals to the network utility, then the global maxima of network utility, i.e., the globally optimal solution to the proposed problem (5) is a pure strategy NE of the proposed EENCG game. Hence, the Theorem is proved.
4. Decentralized Energy Efficient Channel Selection Learning Algorithm
According to above theoretic analysis, a method which obtains the best NE of the game is also the method to achieve the global optimal solution of our problem. In this section, we design a distributed energy efficient channel selection learning (DEECSL) algorithm to obtain the global optimization in a distributed manner.
For potential game, several learning algorithms could achieve the NE points, such as best response dynamic
[28]
, noregret learning
[24]
. However, these learning algorithms may converge to some suboptimal NE points, not the optimal NE point.
In order to escape from suboptimal NE points and finally converge to the optimal NE, the proposed DEECSL algorithm introduces a probabilistic decision mechanism inspired by Boltzmann exploration strategy
[29]
. The DEECSL algorithm is shown in
Table 1
.
The proposed DEECSL Algorithm
The proposed DEECSL Algorithm
In step 4, the learning parameter
β
shows the tradeoff between exploration and exploitation. Smaller
β
implies that players tend to choose a suboptimal action, whereas higher
β
implies that players are prone to choose the best response action. Note that the probabilistic decision in step 4 is designed to avoid the learning algorithm converges to some suboptimal NE points.
Theorem 2:
The proposed DEECSL algorithm will converges to a unique stationary distribution of players’ strategy profile, and the global optimum would be achieved with an arbitrarily high probability with a sufficiently large
β
. The unique stationary distribution of players’ strategy profile
will be given by:
where Γ(
a
) is the potential function.
A
=
A
_{1}
⊗
A
_{2}
⊗...⊗
A_{M}
is the set of strategy profiles of all the players, where ⊗ is the Cartesian product.
Proof:
the following proof is based on the theory in
[29]
and Theorem 6.2 in
[30]
, which is also used in
[18]
,
[19]
,
[23]
. Denote the network state at the tth iteration by
S
(
t
) = (
S
_{1}
(
t
),
S
_{2}
(
t
),...,
S_{m}
(
t
),...,
S_{M}
(
t
)), where
S_{m}
(
t
) means
Ob_{m}
’s working channel in the tth iteration. According to the proposed algorithm,
S
(
t
+ 1) is only determined by
S
(
t
), which means that
S
(
t
) is a discrete time Markov process with a unique stationary distribution
[29]
.
Denote
S
(
t
+ 1) =
a
_{2}
,
S
(
t
) =
a
_{1}
, and denote the transition probability from state
a
_{1}
to
a
_{2}
as
P
_{a1,a2}
, the transition probability from state
a
_{2}
to
a
_{1}
as
P
_{a2,a1}
. There is one or zero element that may be changed between
a
_{2}
and
a
_{1}
in the proposed algorithm (step 3).
When
a
_{2}
=
a
_{1}
, we have:
When
a
_{2}
≠
a
_{1}
, one player changes its working channel, which reduces that one element of the network state has been changed, e.g.,
Ob_{m}
’s working channel from
S_{m}
to
by adopting action
. In step 3, the probability for
Ob_{m}
be has being chosen to update its action is 1 /
M
. Then we have:
Similarly, suppose
Ob_{m}
’s working channel from
S_{m}
to
by adopting action
, we have:
Because the game has been proved be an exactly potential game, we have:
Based on (18), we have:
Thus,
Considering all the states of the network, we have:
According to the discrete time Markov process theory
[29]
, the proposed algorithm has the unique stationary distribution (14) which satisfies the balanced equation of Markov process.
Furthermore, according to
Theorem 1
, the global optimal solution to network utility is exactly the best pure strategy NE of the game. Suppose that
a
^{opt}
is the globally optimal action profile of players, according to the design of the potential function, we have
According to
Theorem 1
, the algorithm converges to a unique stationary distribution
When the
β
→ ∞ , exp {
β
Γ(
a
^{opt}
)} □ exp {
β
Γ(
a
)}, ∀
a
∈{
A
╲
a
^{opt}
}.
Then the probability of globally optimal solution
a
^{opt}
will be
This result means that the proposed learning algorithm converges to the global optimum with an arbitrarily high probability. Thus, the proof is completed.
Notably, the information exchange in the proposed algorithm is strictly constrained between neighboring players, which makes the cost controlled. The global information is not needed for each player. In other words, the proposed algorithm achieves global optimum with only neighboring information exchange.
Furthermore, we analysis the converging speed and the computing complexity of the proposed algorithm, which are two important factors in the practical application.
Theorem 3:
The proposed DEECSL algorithm will asymptotically converge to the global optimum with geometric speed, and the computing complexity would be
O
(
N
).
Proof:
the following proof is based on the theory and approach in
[32]
, also used in
[33]
. According to the above
Theorem 2
, the network state
S
(
t
) is a discrete time Markovprocess with a unique stationary distribution, here denoted by
S
_{π}
. Denote all the possible states of this Markov process as set Ξ
_{S}
.
Denote the initial state as
S
_{I}
in the Initialization step of the proposed algorithm. Then the network state would transmit from
S
_{I}
to
S
_{π}
according to process of learning and updating in the algorithm. Denote the transition matrix of the Markov process as □ , which depict the transition probabilities between the possible network states.
According to
[32]
, define the distance between two states
S
_{1}
,
S
_{2}
as
d
(
S
_{1}
,
S
_{2}
), which measures the variation between
S
_{1}
and
S
_{2}
. Then the distance between
S
_{I}
and
S
_{π}
would be
d
(
S
_{I}
,
S
_{π}
). Construct a matrix ℵ = (
S
_{I}
,
S
_{π}
)
^{T}
, and define ℑ = (
S
_{I}^{T}
□,
S
_{π}^{T}
□)
^{T}
.
The ergodic coefficient
[32]
is defined as follow:
According to the property of ergodic coefficient
[32]
, we have:
Each updating step in the learning algorithm would cause a state transition according to the transition matrix. Suppose the network state is
S_{I}
□
^{k}
after the
k
th state transition after the
k
th learning step. Then the distance between the
k
th state to the stationary state would be:
Furthermore, with the
δ
(□)
^{k}
, according to the definition of ergodic coefficient
[27]
, we have:
where □
_{ij}
is the probability from state
S_{i}
to
S_{j}
.
With the state transition in the proposed algorithm, according to the step 4, the network state transition happens when the selected updating user changes its strategy based on equation (13). Suppose the updating player is
Ob_{m}
, its current strategy is
a_{m}
and the corresponding current network state is
S_{i}
. Denote the possible next strategy is
and the corresponding current network state is
S_{j}
. According to the mixed strategy (13), the network state transition probability from state
S_{i}
to
S_{j}
would be:
where 
A_{m}
 is the number of possible strategies for
Ob_{m}
.
Define
which is biggest gap of utility between all possible two states in the Markov process Ξ
_{S}
. According to the utility function,
is a finited constant value and
≥ 0 . Then we have:
As a result, we have:
Further more:
This means that the algorithm converges from the initial state to the network stationary state geometric speed.
With the computing complexity, in the proposed algorithm, the computing is mainly on the strategy selection, i.e., for the seleceted strategy updating player,
Ob_{m}
, the probabilities for possible strategies need to be computed by:
There are 
A_{m}
 ≤ 
S_{CH}
 =
N
possible strategies for
Ob_{m}
to be computed. Then the computing complexity would be
O
(
N
). Thus, the proof is completed.
It should be noted that geometric converging speed should be acceptable for most applications, and the computing complexity is much smaller than the central control based exhausive searching approach, whose computing complexity would be
O
(
N^{M}
).
5. Numeric Results and Discussion
In this section, we evaluate the performance of the proposed algorithm by Matlab simulations. Without loss of generality, the following parameters are set: the number of channels is 5. Similar with
[31]
, we set the path loss exponent as
γ
= 2, and the noise power as
N
_{0}
= −130 dB. The transmission distance for Obs is the unit distance for simplicity. The transmission data rate is 1 Mbit/s. The instantaneous random components are assumed to be unitconstant and the channels are assumed to undergo Rayleigh fading with unit mean. The neighboring relationship of
Ob
s is randomly generated. The bandwidth of channels are set as [1, 1.5, 2, 2.5, 3] MHz to simulate the different channels. The results are obtained by simulating 3000 independent experiments and then taking the average value.
Simulation Parameters
The simulation is to show the performance of the proposed DEECSL algorithm compared with other approach, and verify the theoristic analysis of proposed algorithm, i.e., the convergence and the optimality. We first verify the convergence and the optimality of the proposed algorithm in a relative small scenario. Then we show the performance comparison furthermore in a relative large scenario to verify the adaption of the proposed algorithm.
To show the convergence and the optimality of the proposed algorithm, we compare the performance of the proposed algorithm with the globally optimal solution obtained by exhaustive search. We simulate 15 players scenario to avoid the limitation of computing capacity with the exhaustive search method.
Fig. 2
shows that the proposed DEECSL algorithm converges to the globally optimal network utility after some learning iterations. This result verifies the theoristic analysis on the proposed approach.
The convergence and optimality of the proposed DEECSL algorithm, M=15.
To show the advantage of the collaboration design in the proposed DEECSL algorithm, we also show the performance of the traditional noncollaboration approach, which is widely adopted in game theoristic learning algorithms. In the noncollaboration approach, each player does not care other neighboring players’ benefits and make the decision totally based on its own reward. The result shows that the proposed collaboration design DEECSL algorithm outperforms the noncollaboration approach obviously. The main reason is that players in the noncollaboration approach would not consider the golobal reward, that also results in the unstability of noncollaboration approach. This comparison result verifies necessity and the advantage of the collaboration based utility function design in the proposed game model. In other words, with out the utility function design, the system would not converge to stable state and the performance would be worse.
In order to show the advangate of the design of probabilistic decision in the proposed algorithm, we compare with the best response (BR) algorithm
[28]
.
Fig. 2
also shows that the proposed DEECSL algorithm obtains better performance compared with the BR algorithm. Here the BR algorithm is also based on the collaboration approach. The BR algorithm is a typical learning algorithm for games to achieve the NE points. The differenc between the BR algorithm and the proposed DEECSL algorithm is that players in BR algorithm choose the best strategy from all the candidate strategies rather than the probabilistic decision according to equation (13). Due to the “choosing the best” strategy, the BR algorithm’s performance increases faster in the beginning and converges earlier. However, because players choose the current best strategies absolutely, they might loose some potential strategies in the long term. As a result, the BR learning algorithm may converges to some local optimum, which reduces the converged performance.
Compared with the BR algorithm, the proposed algorithm could escape from local optimum to achieve globally optimum due to the design of probabilistic decision. In other words, the probabilistic decision based on equation (13) rather than the “absolutely choosing the current best” would bring the chance for the player to make the right decision in the long term.
To verify the proposed algorithm furthermore, we simulate a larger network scale where the number of players is 50. Due to the limitation of computation capability, the globally optimal solution by the exhaustive search is not given out. We compare the proposed algorithm with the BR learning algorithm and the noncollaboration approach.
Fig. 3
shows the convergence of the proposed DEECSL algorithm again. Also, the proposed DEECSL algorithm achieves better performance than the BR algorithm. The noncollaboration approach achieves worst performance and unstability for the lack of collaboration. This result verifies the theoristic analysis of the proposed algorithm and the performance advantage compared with other approaches again. That means the proposed approach is widely applicative in different scenarios.
The proposed DEECSL algorithm vs the BR learning algorithm, M=50.
6. Conclusion
In this paper, we studied the distributed energy efficient resource management in the Internet of Things. We proposed an energy efficient neighbor collaborative game model and proved that the proposed game was an exact potential game. Then, to achieve the global optimization to the proposed energy efficient channel selection problem in a distributed manner, we designed a distributed energy efficient channel selection learning algorithm to obtain the global optimization with only neighboring information exchange. We proved that the proposed DEECSL algorithm will asymptotically converge to the global optimum with geometric speed. Simulation results verified the theoretic analysis and the performance of proposed algorithm.
BIO
Chunhua Ju graduated from Xiamen University with PhD degree in 2002. Now he is a professor,doctoral supervisor and division chief of science and technology department in Zhejiang Gongshang University. He focuses on intelligent information processing,data mining and collaborative innovation. He won the award for“New Century Excellent Talentsin University”of China.
Qi Shao received the B.E. degree in Computer Science and Technology from Zhejiang Gongshang University, China, in 2014. He is currently pursuing the M.A. degree in Computer Science and Technology in Gongshang University, China.
Wu Q.
,
Ding G.
,
Xu Y.
,
Feng S.
,
Du Z.
,
Wang J.
2014
“Cognitive Internet of Things: A New Paradigm Beyond Connection,”
IEEE Internet of Things Journal
1
(2)
DOI : 10.1109/JIOT.2014.2311513
Agrawal S.
,
Das M. L.
2011
“Internet of Things – A Paradigm Shift of Future Internet Applications,”
Institute of Technology, Nirma University
AHMEDABAD
Lin F.
,
Liu Q.
,
Zhou X.
,
Chen Y.
,
Huang D.
“Cooperative Differential Game for Model EnergyBandwidth Efficiency Tradeoff in the Internet of Things,”
Network Technology and Application, China Communications
Jan. 2014
Zheng X.
,
Martin P.
,
Brohman K.
,
Xu L.
2014
“Cloud Service Negotiation in Internet of Things Environment: A Mixed Approach,”
IEEE Transactions on Industrial Informatics
10
(2)
Serbulent T.
,
Murat S.
,
Wei M.
2012
“WiFi Enabled Sensors for Internet of Things: A Practical Approach,”
IEEE Communications Magazine
50
(6)
134 
143
DOI : 10.1109/MCOM.2012.6211498
Sanchez L. T.
,
Ranasinghe D. C.
,
Harrison M.
2012
“Adding Sense to the Internet of Things,”
Personal and Ubiquitous Computing
16
(3)
291 
308
DOI : 10.1007/s0077901103998
Roman R.
,
Alcaraz C.
,
Lopez J.
2011
“Key Management Systems for Sensor Networks in the Context of the Internet of Things,”
Computers and Electrical Engineering
37
(2)
147 
159
DOI : 10.1016/j.compeleceng.2011.01.009
Klauck R.
,
Kirsche M.
“A Case Study of a DNSBased Discovery Service for the Internet of Things,”
ADHOCNOW
2012
316 
329
Chen Z.
,
Xia F.
,
Huang T.
2011
“A Localization Method for the Internet of Things,”
Journal of Supercomputing
1 
18
Wu T. Y.
,
Li G. H.
,
Huang S. W.
2012
“A GABased Mobile RFID Localization Scheme for Internet of Things,”
Personal and Ubiquitous Computing
16
(3)
245 
258
DOI : 10.1007/s0077901103989
Chen Z.
,
Wang H.
,
Liu Y.
2012
“A ContextAware Routing Protocol on Internet of Things Based on Sea Computing Model,”
Journal of Computers
7
(1)
96 
105
DOI : 10.4304/jcp.7.1.96105
Danieletto M.
,
Bui N.
,
Zorzi M.
“Improving Internet of Things Communications Through Compression and Classification,”
in Proc. of the 2012 IEEE International Conference on Pervasive Computing and Communications Workshops
2012
Li S.
,
Fan C.
2011
“Joint Optimization on Energy and Delay for Target Tracking in Internet of Things,”
China Communications
8
(1)
20 
27
Duan J.
,
Gao D.
,
Yang D.
,
Foh C. H.
,
Chen H.
2014
“An EnergyAware Trust Derivation Scheme With Game Theoretic Approach in Wireless Sensor Networks for IoT Applications,”
IEEE Internet of Things Journal
1
(1)
DOI : 10.1109/JIOT.2014.2314132
Fudenberg D.
,
Levine D. K.
1998
“The Theory of Learning in Games,”
The MIT Press
Suris J.
,
Dasilva L.
,
Han Z.
,
Mackenzie A.
,
Komali R.
2009
“Asymptotic optimality for distributed spectrum sharing using bargaining solutions,"
IEEE Transactions on Wireless Communications
8
(10)
5225 
5237
DOI : 10.1109/TWC.2009.081340
Zhang Z.
2008
“A cooperation strategy based on Nash bargaining solution in cooperative relay networks,"
IEEE Transactions on Vehicular Technology
57
(4)
2570 
2577
DOI : 10.1109/TVT.2007.912960
Zheng J.
2014
“Optimal power allocation and user scheduling in multicell networks: Base station cooperation using a gametheoretic approach,”
IEEE Trans. Wireless Commun
13
(12)
6928 
6942
DOI : 10.1109/TWC.2014.2334673
Xu Y.
,
Wang J.
,
Wu Q.
,
Anpalagan A.
,
Yao Y.D.
2012
“Opportunistic spectrum access in cognitive radio networks: Global optimization using local interaction games,”
IEEE J. Sel. Signal Process
6
(2)
180 
194
DOI : 10.1109/JSTSP.2011.2176916
Dai H.
,
Huang Y.
,
Yang L.
2015
“Game theoretic maxlogit learning approaches for Joint Base Station Selection and Resource Allocation in Heterogeneous Networks,”
IEEE Journal on Selected Areas in Communications
33
(6)
DOI : 10.1109/JSAC.2015.2416988
Montanari A.
,
Saberi A.
“Convergence to equilibrium in local interaction games,”
in Proc. 50th Annu. IEEE Symp. Foundat. Comput
2009
303 
312
Stuber G.
2001
Principles of Mobile Communications
2nd ed.
Kluwer
Norwell, MA, USA
Marden J.
,
Arslan G.
,
Shamma J.
2009
“Cooperative control and potential games,”
IEEE Trans. Syst., Man, Cybern. B
39
(6)
1393 
1407
DOI : 10.1109/TSMCB.2009.2017273
Maskery M.
,
Krishnamurthy V.
,
Zhao Q.
2009
“Decentralized dynamic spectrum access for cognitive radios: cooperative design of a noncooperative game,”
IEEE Trans. Commun
57
(2)
459 
469
DOI : 10.1109/TCOMM.2009.02.070158
Santos F. C.
,
Santos M. D.
,
Pacheco J. M.
2008
“Social diversity promotes the emergence of cooperation in public goods games,”
Nature
454
213 
249
DOI : 10.1038/nature06940
Altman E.
,
Jimenez T.
,
Vicuna N.
,
Marquez R.
“Coordination games over collision channels,”
in Proc. WiOPT
2008
523 
527
Zhong W.
,
Xu Y.
,
Tianfield H.
2011
“Gametheoretic opportunistic spectrum sharing strategy selection for cognitive MIMO multiple access channels,”
IEEE Transactions on Signal Processing
59
(6)
2745 
2759
DOI : 10.1109/TSP.2011.2121063
van Laarhoven P. J. M.
,
Aarts E. H. L.
1987
Simulated Annealing: Theory and Applications
Reidel
Holland
Young H. P.
1998
Individual Strategy and Social Structure
Princeton Univ. Press
Princeton, NJ
Babadi B.
,
Tarokh V.
2010
“GADIA: A greedy asynchronous distributed interference avoidance algorithm,”
IEEE Trans. Inf. Theory
56
(12)
6228 
6252
DOI : 10.1109/TIT.2010.2081090
Brémaud P.
1999
Markov Chains: Gibbs Fields, Monte Carlo Simulation, Queues.
SpringerVerlag
New York, NY, USA
Li Y.
,
Jin D.
,
Yuan J.
,
Han Z.
2014
“Coalitional Games for Resource Allocation in the DevicetoDevice Uplink Underlaying Cellular Networks,”
IEEE Transactions on wireless communications
13
(7)