Advertisement

Advertisement

Applications of game theory in deep learning: a survey

  • Published: 09 February 2022
  • Volume 81 , pages 8963–8994, ( 2022 )

Cite this article

research paper about game theory

  • Tanmoy Hazra 1 &
  • Kushal Anjaria 2  

29k Accesses

29 Citations

7 Altmetric

Explore all metrics

This paper provides a comprehensive overview of the applications of game theory in deep learning. Today, deep learning is a fast-evolving area for research in the domain of artificial intelligence. Alternatively, game theory has been showing its multi-dimensional applications in the last few decades. The application of game theory to deep learning includes another dimension in research. Game theory helps to model or solve various deep learning-based problems. Existing research contributions demonstrate that game theory is a potential approach to improve results in deep learning models. The design of deep learning models often involves a game-theoretic approach. Most of the classification problems which popularly employ a deep learning approach can be seen as a Stackelberg game. Generative Adversarial Network (GAN) is a deep learning architecture that has gained popularity in solving complex computer vision problems. GANs have their roots in game theory. The training of the generators and discriminators in GANs is essentially a two-player zero-sum game that allows the model to learn complex functions. This paper will give researchers an extensive account of significant contributions which have taken place in deep learning using game-theoretic concepts thus, giving a clear insight, challenges, and future directions. The current study also details various real-time applications of existing literature, valuable datasets in the field, and the popularity of this research area in recent years of publications and citations.

Similar content being viewed by others

research paper about game theory

Games of GANs: game-theoretical models for generative adversarial networks

research paper about game theory

A Deep Analysis on the Role of Deep Learning Models Using Generative Adversarial Networks

research paper about game theory

Understanding GANs: fundamentals, variants, training challenges, applications, and open problems

Explore related subjects.

  • Artificial Intelligence

Avoid common mistakes on your manuscript.

1 Introduction

Game theory is an essential field for research, and it helps to choose the suitable strategy of the players in a game. It has several applications in various domains. Game theory has significant applications in technology, especially concerning computer science, electronics, aerospace engineering, etc. [ 90 , 91 , 96 , 99 , 100 ]. Different types of games are briefed in Fig.  1 . Alternatively, Deep Learning is the study of various learning algorithms that uses multiple layers of non-linear processing units. The output of the previous layer is taken as input by each successive layer. Deep learning algorithms are primarily classified into three categories, e.g., supervised, semi-supervised and unsupervised. Deep Learning algorithms implement higher-level features that are extracted from lower-level features. The depth in neural networks plays an essential part in the outcome of the model. The framework of the neural networks can represent dynamic environments, and similarly, dynamic environments can further be presented as games [ 37 , 72 ].

figure 1

Types of games

Artificial intelligence has adapted the game theory to solve or model various real-time problems, and it is observed that the performance is improved while applying game theory [ 43 , 76 , 124 , 148 ]. This paper establishes the connection of deep learning algorithms with game theory. The applications of evolutionary computing and swarm intelligence are discussed in [ 20 ]. This paper mainly focuses on applications of game theory to solve GAN models. GAN has received a tremendous response from several research communities because of its complex problem-solving abilities and performance improvement [ 19 , 36 , 130 ].

The multilayer neural network comprises any number of unit neurons and may determine multimodal output functions. The multimodality nature of the output function creates hardships to optimize deep neural network models. Strategic Deep Learning is a defying game task [ 62 , 84 , 136 , 141 , 151 ]. Table 1 shows research articles in game theory, deep learning, and their collaborative research as per the records obtained from various databases. In a nutshell, the contribution of the paper is as follows:

This paper reviews several significant contributions to applying game theory in deep learning models and achieving performance improvement.

Paper considers a variety of evolutionary algorithms from game-theoretic perspectives.

The paper suggests how the bonding between game theory and deep learning can be strengthened.

Indexing of this paper is as follows: basics and background of game theory and deep learning are discussed in section 2; game theory in deep learning and artificial intelligence is discussed in section 3; after that, game theory in reinforcement learning (section 4) is followed by game theory in GAN (section 5); section 6 discusses implementation environment of the current study; section 7 outlined challenges and future direction and section 8 summarizes the discussion.

2 Basics and background

This section discusses fundamentals and rudimentary concepts that are directly or indirectly applied in the existing literature. This section is a platform for the readers to enter into the field and helps readers understand advancements in the field. The objective of this section is to formally introduce important terms in both areas to the readers before elaborating their connections in subsequent sections. This paper discusses various applications of game theory in the domain of deep learning. Thus, this section initially introduces the game theory, followed by an introduction to deep learning architectures and techniques. In the subsequent sections, the scope of the paper is narrowed down, and specific areas of game theory and deep learning are discussed in detail. Section 2.1 explains various game-theoretic concepts, and section 2.2 discusses multiple deep learning architectures and techniques.

2.1 Game theory overview

Game theory is a significant area of applied mathematics to model various real-world complex scenarios as games. Thus, game theory is broadly used in multiple fields where strategic interaction among the players plays a vital role.

A game can be defined as a competitive or cooperative interaction among the objects or players [ 89 ]. The nature of interactions among players determines the behavior of games. There are several types of games in the game theory: cooperative and non-cooperative, zero and non-zero, sequential and simultaneous, normal-form and extensive-form, finite and infinite, perfect-information and imperfect-information combinatorial, Bayesian, stochastic, etc. In this paper, cooperative and non-cooperative, zero and non-zero, normal-form and extensive-form, perfect-information and imperfect-information, finite and infinite, and stochastic games are addressed, and the games are played between single or multiple players. The basic concepts of the games mentioned above are as follows:

Cooperative and non-cooperative games: The players cooperatively choose their strategies to maximize the overall payoff [ 23 ]. Alternatively, the players choose their strategies to maximize their self-payoffs for non-cooperative games [ 93 ].

Zero and non-zero games: In zero-games, the value of a game is zero. On the other hand, it can be defined as the summation of payoffs for adversary players for any set of actions must be zero. Alternatively, the value of a non-zero game is non-zero [ 126 ].

Normal-form and extensive-form games: Normal-form games are represented by matrices that include possible strategies for players in the topmost row and leftmost column. The payoffs of players are represented as matrix elements separated by commas. Alternatively, extensive-form games are described as trees, where internal nodes represent players’ turns, edges represent players’ actions, and external nodes represent players’ outcomes for a particular set of moves. Figure 2a and b illustrate examples of the normal-form and extensive-form game representations [ 79 ].

Perfect-information and imperfect-information games: Perfect information games can be defined as a game where players have the same information. In other words, players are aware of all actions that have occurred in the past. In imperfect information games, players choose their moves simultaneously. So, they are not perfectly informed about the moves of opponents [ 88 ].

Finite and infinite games: Finite games/finite strategy games are the games with finite strategies. Alternatively, infinite games are games with infinite strategies [ 12 ].

Stochastic games: The game is played in a sequence of time units/stages. In each stage, games enter into a new state. Payoffs of players depend on the present state and chosen action [ 117 ].

figure 2

Game representation ( a ) Normal-form ( b ) Extensive-form

2.1.2 Players

Players are the critical components of a game. They participate in a game intending to achieve certain goals. The nature of interaction among the players can be primarily classified into two types: cooperative and non-cooperative. Two or more than two players can take part in a game. The players always aim to maximize the overall payoff of the whole group.

2.1.3 Strategies

In a game, strategies are a sequence of actions chosen by the players. The strategies can also be fundamentally classified into two types, cooperative and non-cooperative, based on players’ goals. The paramount of a game is to determine suitable strategies for players.

2.1.4 Payoffs

The payoffs of players represent the rewards or penalties for choosing their respective strategies. In cooperative games, players aim to improve the overall payoffs of a group. In contrast, the players aim to maximize their self-payoff (without considering the overall payoff of the group) in non-cooperative games. The game planners formulate the payoff functions. Let St I represents all possible strategies for player I . If there are m number of players, then, possible combinations of strategies for player I is St 1  ×  St 2 … St m . So, the payoff function can be represented as util ( St 1  ×  St 2 … St m ).

2.1.5 Best response

Best response strategies denote the most favorable outcomes for a player considering all the strategies for opponents. The players in a game usually prefer to choose the best response strategies. NE is determined based on the best response strategies for players. Let best response strategy for player I be represented as br I (.). Thus, considering a set of opponent’s strategies x − I , br I ( x − I ) denotes player I ’s best response to x − I .

2.1.6 Nash equilibrium

It is a stable state where players cannot earn more profits by unilaterally deviating from their strategies. Players can have a pure or mixed strategy equilibrium in a game.

Let a game with m players be denoted by ( St ,  util ), St I represents strategies for player I , St  =  St 1  × … ×  St m represents strategy combinations and util ( x ) = ( util 1 ( x ), …,  util m ( x )) represents the utility or payoff function, where xϵSt . x I and x − I are the strategy combination for player I and others except for player I , respectively. When each player Iϵ 1, …, m computes the strategy x I from the strategy-profile x  =  x 1 , …, x m , then, the player I gets a payoff util I ( x ). A strategy profile x ∗ ϵSt is a NE if the following condition is satisfied

Mixed equilibrium is computed for those scenarios, where the players mix their strategies with uncertainties, and the mixed equilibrium is computed based on the expected payoffs of respective players [ 92 ].

2.1.7 Shapley function

In cooperative games, the reward of each player is computed by a function called Shapleyfunction. The Shapley value \({\Phi}_{S_i}(v)\) for each player S i , 1 ≤  i  ≤  N is computed based on its individual contribution to a coalition, where N represents the player count. The Shapley value [ 116 ] for each player is calculated by the equation given below.

In the above equation, S represents a set of N players (| S | =  N ), C denotes a group (| C | =  c ) and C is a subset of S\{ S i }, \(\frac{\mathrm{c}!\left(\mathrm{N}-\mathrm{c}-1\right)}{\mathrm{N}!}\) denotes the uncertainty in a permutation, the contributors of C are ahead of the individual player S i and (v(C ∪ { S i }) − v(C)) denotes the individual contribution of a player S i in the group C , where the rewards of all groups are initially determined. In this way, the Shapley function \({\Phi}_{S_i}(v)\) yields the Shapley value of a player S i for a particular group. A player having the highest Shapley value is the most significant contributor in a group.

2.1.8 Minimax theorem

In 1928, John Von Neumann introduced the Minimax theorem, which opens the door for conventional game theory. For a two-player, zero-sum, simultaneous move finite game, there must be a value and exists an equilibrium point for both the players [ 126 ]. The equilibrium point can be determined by applying pure or mixed strategies by either one or both the players. Let’s assume that x and y are strategies of two players and v is the value of the game. Then, the minimax theorem can be formulated as

2.2 Deep learning overview

Deep learning is an advanced part of machine learning algorithms based on artificial neural networks and various learning methods, i.e., supervised, unsupervised, and reinforcement. There are well-known deep learning architecture and techniques. Deep learning models use multiple layers in an artificial neural network to extract significant characteristics from the unprocessed data. Graphics Processing Units (GPUs) are required to perform high-power computation on complex deep learning architectures. Deep learning has immense applications in multi-dimensional fields such as speech recognition, computer vision, audio recognition, natural language processing, medical image analysis, games, etc.

2.2.1 Deep learning architecture

In recent times, deep learning has received tremendous responses from various fields. Robust deep learning architectures bring significant improvement in the performance of multiple models. They can solve or model various complex problems because of their robustness. In most cases, it is observed that deep learning architectures outperform other existing models. The most popular deep learning architectures are the convolutional neural network, recurrent neural network, generative adversarial network, deep belief network, autoencoder, residual neural network, etc. This paper primarily studies generative adversarial neural networks because of their architecture, where game-theoretic techniques can be easily applied. In the future, we may explore possibilities of applications of game theory in other deep learning architectures.

Convolutional neural network

The convolutional neural network, a significant type of deep learning architecture, is best known for its vast capabilities when analyzing visual imagery. They comprise regularized versions of complicated multilayer perceptrons (usually fully connected networks). A CNN has three layers, input, hidden, and output. The hidden layers are mainly convolutional layers that convolve with multiplication or other product, and the activation function is commonly a RELU layer. CNNs were inspired by the works [ 32 , 59 , 60 ].

Recurrent neural network

Recurrent Neural Network is a category of advanced artificial neural networks. A directed graph with a temporal sequence is what we get when there is a connection between the nodes. Temporal dynamic behavior is exhibited in RNN. Using internal states, RNNs have capabilities to process inconsistent range sequences of inputs. They are powerful enough because having a distributed hidden state makes them capable of storing huge amounts of information regarding the past and non-linear dynamics, acknowledging them to revise their hidden state in complex ways [ 39 , 111 , 146 ].

Long Short-Term Memory, applied in deep learning, belongs to the family of recurrent neural networks. Its specialty lies in the fact that along with feed-forward neural networks, it has feedback connections too. A general LSTM architecture constitutes a cell, an input gate, an output gate, and a front gate. The utility of a cell is to store values over irregular periods, and all gates examine the flow of data entering into and exiting out of the cell. LSTM networks are best applicable in categorizing, processing, and predicting based on time series data [ 15 ,  54 ].

Generative adversarial network

Generative adversarial network (GAN), a type of deep learning architecture, was invented in 2014 by Ian Goodfellow [ 36 ]. Here, two neural networks challenge amongst themselves in a game (initially given a training set). The theory of GAN is to train images that can generate new images that look accurate to human eyes. GANs were originally modeled for unsupervised learning, but they are broadly used in reinforcement learning, semi-supervised learning, and fully supervised learning.

GAN layered architecture is represented in Fig. 3 . Let’s introduce the variables and parameters related to the GAN model

Discriminator

Parameters of the discriminator

Parameters of generator

Input noise distribution

Original data distribution

Generated distribution

figure 3

The layered architecture of GAN

The final loss function for the discriminator can be written as

Alternatively, the generator is competing with the discriminator. The final loss function for the generator can be written as

Both the equations can be combined and rewritten as

The above equation represents a single data point. For an entire dataset, we can write the equation as follows:

Deep belief networks

A deep belief network, a type of deep learning architecture, is a probabilistic generative graphical model. They are constructed from several layers of stochastic, latent variables with binary values called hidden layers or feature identifiers. The two compelling features of deep neural networks are:

A sequential layer-wise, efficient method for learning the top to bottom, generative weights capable of determining how variables on a layer depend upon the variables on its top layer

After learning, values are determined for latent variables in each layer, resulting in the single, bottom-up pass. It starts with a vector of selected data in the bottom layer. It exploits the weights in the opposite direction [ 48 , 49 , 50 ].

Autoencoder

Autoencoder, a distinctive category of artificial neural network utilized in learning efficient data coding through unsupervised techniques. Autoencoders target to learn an encoding for a set of given data and to reduce dimensions in data by training the given network to ignore “noise.” They can probably encode a provided input to represent smaller dimensions, being a data-compression model. Later, decoders can then reconstruct the input back from the encoded version [ 125 , 147 ].

Residual neural network

Residual neural networks, abbreviated as ResNet, are artificial neural networks of a particular type inspired by the pyramidal cells of the animal cerebral cortex. Residual neural networks perform this by applying skip connections or shortcuts to jump through a few layers. ResNet models are achieved using double/triple layer skips consisting of nonlinearities (ReLU) and batch normalization in between [ 46 ]. An extra new weight matrix may be used to learn the skip weights. Such architectures are known as HighwayNets [ 121 ]. Architecture with several parallel skips is termed as DenseNets [ 57 ].

Radial basis function networks (RBFNs)

“The RBF network model is motivated by the locally tuned response observed in biologic neurons. Neurons with a locally tuned response characteristic can be found in several parts of the nervous system, such as cells in the auditory system that are selective to small bands of frequencies [ 114 ]”.

Multilayer Perceptrons (MLPs)

“MLPs belong to the class of feedforward neural networks with multiple layers of perceptrons that have activation functions. MLPs consist of an input layer and an output layer that is fully connected. They have the same number of input and output layers but may have multiple hidden layers and can be used to build speech recognition, image recognition, and machine-translation software [ 8 ].”

Self-organizing maps (SOMs)

“The SOM algorithm distinguishes two stages: the competitive stage and the cooperative stage. In the first stage, the best matching neuron is selected. In the second stage, neuron weights are not modified independently but as topologically-related subsets on which similar kinds of weight updates are performed [ 66 ].”

2.2.2 Deep learning techniques

Dropout is adapted as a technique to overcome overfitting problems in a neural network. It addresses both issues – training and testing computations. Effectively, it allows the training of several neural networks without any significant computational overhead. Also, it gives an efficient approximate way of combining exponentially many different neural networks. In the training phase, dropping out refers to dropping out units(neurons) of a certain set of randomly chosen neurons. The dropped out units are not further considered during a forward and backward pass. Temporarily removes a node and all its incoming/outgoing connections, resulting in a thinned network [ 49 , 50 ].

Rectified linear unit

ReLU is an activation function broadly used in various deep learning architectures. It is a non-linear activation function used for both types of networks, i.e., multiple-layer neural networks and deep neural networks. The function for any negative input returns zero. Alternatively, x is returned by the function for any positive input x . So, the simplified form of the function is f ( x ) = max(0,  x ).

In recent times, a sigmoid and hyperbolic tangent is replaced by the ReLU function. The main reason for the popularity of the ReLU function is its ability to make the training speed of deep neural networks faster than other conventional activation functions. A significant feature of ReLU is the derivative of this function is 1 for positive input. Deep neural networks can save additional time for calculating error terms due to a constant during the training phase. The extensive use of ReLU is shown in [ 34 ].

Stochastic gradient descent

Stochastic gradient descent is one of the powerful algorithms used in several machine learning and deep learning models. It is the basis of neural networks. It is an essential iterative algorithm. The functionalities of the gradient descent algorithm are as follows: it starts from an arbitrary point on a function. It moves down its slope in several steps/iterations until it reaches the minimum point of the given function. The algorithm is modified by including a random probability, called Stochastic Gradient Descent. In this algorithm, a set of samples is randomly chosen from the whole data set in each iteration. If we consider a large dataset, the programmer may require using many samples in each iteration while using the Gradient Descent algorithm. The task needs to be performed for each iteration until the minima are found, which is the main challenge in the Gradient Descent algorithm. Computational complexity is a significant concern in this algorithm. Stochastic Gradient Descent is introduced as a solution to this problem. SGD reduces the sample size. It takes only a single sample to perform the task for each iteration. Therefore, the sample is randomly mixed and chosen for completing the task for the iteration. The backbone of the SGD is to consider the gradient of the cost function of a single sample for each iteration. Applications of the Gradient Descent method are addressed in [ 26 , 65 ].

Batch normalization

Batch normalization is an essential technique in artificial neural networks. The advantages of the method are improvement in speed, stability, and performance of neural networks. The input layer can be normalized by scaling and adjusting the activation functions. It is a technique by which the inputs for each layer are normalized to deal with the internal covariate, shift problem, i.e., the problem appears in the intermediate layers because, during training, the distribution of the activation functions is constantly changing. This change slows down the training process because each layer needs to learn a new distribution of activation functions in each training step. This method includes: calculating the variance and mean of the layer inputs, normalizing the layer inputs with the help of batch statistics, scaling, and shifting to find the layer’s output [ 61 ].

3 Application of game theory in deep learning and artificial intelligence

This section addresses some contributions in which game models solve deep learning and artificial intelligence problems. Figure 4 depicts the basic architectures of a simple neural network and deep learning neural network.

figure 4

( a ) Simple Neural Network ( b ) Deep Learning Neural Network

The authors in [ 112 ] address a new approach by which game-theoretic techniques can model individual neurons. The authors show that different strategic game-theoretic approaches can be applied to model paired neuron systems. A learning algorithm is developed depending on game theory for neural learning. Artificial neural network has proved its significance in multi-dimensional domains. Selecting an appropriate network is challenging to solve a problem [ 62 , 141 , 151 ]. The authors in [ 122 ] show that game-theoretic concepts like the Shapley value help to differentiate significantly from unnecessary elements of an artificial neural network. A cooperative game is designed from a neural network where neurons that form different groups and their contributions to the game are determined with the help of the Shapley value. The experiments prove that the Shapley value concept is better than other heuristics approaches to assess the contribution of neurons.

There are various GAN algorithms. The first is fully connected neural networks for both the generator and discriminator. The second is convolutional GAN, as going from fully connected to convolutional neural networks is a natural extension. The third neural network is conditional GAN. It extends the 2D GAN framework to the conditional setting by making both the generator and the discriminator networks class-conditional [ 142 ].

Alternatively, a max-min problem is formulated for adversarial learning with multiplayer stochastic games and two-player sequential games above deep learning networks [ 17 ]. The experimental results forecast the efficacy of the used adversarial algorithm. The algorithm can manipulate adversarial features that affect testing results in deep learning models. The work introduces a secure learner who is adaptive to the antagonistic attacks on deep learning. The paper claims that the given framework is more robust than a traditional convolutional neural network (CNN) and a generative adversarial network under adversarial attacks. This paper also highlights the impacts on adversarial payoff functions over randomized strategies while the rules of the games are changed. The offensive scenarios over such strategy spaces find multiplayer games over varied strategies. A reduction of supervised learning to the game is explored in [ 113 ]. For convex one-layer problems, an equivalence between Nash equilibria in a simple game and global minimizers of the training problem is shown. It is also demonstrated how the game can be extended to acyclic neural networks using differentiable convex gates. The work in [ 86 ] presents a model that integrates the concepts from the field of deep learning and artificial life to reflect their potentiality in various scenarios. The model shows the potentiality of neural networks to simulate population dynamics. The model also shows the applications of evolutionary game theory result in the behavior of the networks.

Modeling humans’ ability to depict the mental state transitions of others is a challenging task for the research community. The authors of [ 106 ] train a machine to construct such models. A Theory of Mind neural network (ToMnet) is designed by meta-learning that builds handler models by observing their behavior. The ToMnet model is applied to agents in simple grid environments. This system can autonomously learn modeling other agents, which is a significant contribution to designing multi-agent AI systems. It can be applied to develop technologies for machine-human communication and advance the growth of interpretable artificial intelligence. For large-scale perfect-information games, artificial intelligence is superior to human-level intelligence [ 124 ]. On the other hand, it is not easy to find good results in large-scale imperfect-information games (i.e., business strategies, war games, etc.). NFSP is a self-play-based approach without prior information, which helps for effectively learning approximate Nash equilibrium. But, the algorithm depends on Deep Q-Network. In online games, it is not easy to converge by changing opponents’ strategies. Neither in large search scale nor deep search depth games can it find approximate Nash equilibrium. This paper introduces the Monte Carlo Neural Fictitious Self Play (MC-NFSP) algorithm that amalgamates the NFSP and Monte Carlo tree search. For large-scale zero-sum imperfect-information games, the approach improves performance. The asynchronous Neural Fictitious Self Play (ANFSP) model is developed to use a parallel and asynchronous framework to gather the game’s history [ 145 ].

The authors in [ 128 ] address a reversed reinforcement learning method. After training a deep neural network according to strategies in the payoff table, randomized strategy input is initialized, and the error differentiates the actual output. The required output is propagated back to the initially randomized strategy input in the input layer of the trained deep neural network results in performing a task similar to the human deduction. Detecting imaging biomarkers for autism spectrum disorder (ASD) is challenging to help to explain ASD and predicting or monitoring treatment outcomes. Deep learning classifiers are used to detect ASD from functional magnetic resonance imaging (fMRI) with better correctness than traditional learning strategies. The concept of Shapley value from cooperative game theory is applied to this problem. Cooperative game theory is suitable since it more accurately determines biomarker importance for each instance from deep learning models. The main challenge for using Shapley Value calculation is its computational complexity. The method is validated on the MNIST dataset and compared to human perception. A Random Forest (RF) is modeled is trained to classify ASD or control subjects from fMRI and compare Shapley value outcomes with existing RF-based feature importance [ 74 ]. The development of intelligent machine learning applications is studied with expected-long-term profit maximization in multi-agent systems. A learning algorithm for the IPD problem is proposed in [ 2 ]. It is shown to outperform the tit-for-tat algorithm and many other adaptive and non-adaptive strategies using numerical analysis. It is also discussed how artificial intelligence and machine learning work closely to provide the agent with a mind-reading capability.

Shin et al. [ 119 ] have put forward a model to assign the updating time period to the drones by auctioning. Alternatively, a second-price auction system is applied in which the victorious bidder pays the second-highest bid. In the model, data needed for the dispersal of drones was found from the deep learning algorithm. The shortcoming of the model proposed by Shin and the team was it does not consider two significant criteria; the possibility of increasing the charging station and the uncertainty of charging to the drones at smaller bids. Ren et al. [ 107 ] have summarized the representative defenses developed, including adversarial training, randomization-based schemes, de-noising methods, and provable defenses.

Further, to utilize the adversarial training-based intuitive training method, Ren et al. have combined min-max games with deep learning and neural network. Leckie et al. [ 71 ] have combined game theory and deep learning to evade jamming attacks in network security. Based on a deep analysis of node behavior characteristics in the opportunistic network, Wang et al. [ 135 ] have introduced the evolutionary game theory to traverse the node cooperation mechanism in the opportunistic social network. In the same line, Ranadheera et al. [ 118 ] have utilized a deep learning-based deep-Q algorithm with the game theory for fair and efficient resource management in mobile edge computing. To achieve intelligence in the shared environment with multiple agents, Lanctot et al. Further, Lu and Kai [ 78 ] have generalized a multi-agent approach using reinforced deep learning and game theory. Dasgupta et al. [ 21 ] have combined deep learning and game theory for cybersecurity approaches.

Rudral et al. [ 108 ] have used deep learning-based predictive modeling for football games to result in multilayer perception. Wakatsuki et al. (2020) have used multi-player games for decision making. Wang et al. [ 134 ] have used the Deep Neural Network (DNN) based deep learning model and the game theory to provide a holistic framework of robot-human interaction. Pinto et al. [ 104 ] have used self-supervised deep learning for the robotic adversary. Moreover, they have designed adversarial training as a two-player zero-sum repeated game. Balduzzi [ 22 ] has used gradient-based game theory and deep learning to optimize game grammars. Shu et al. [ 44 ] have used DNN and game-theoretical approach for developing Multi-granularity Network Representation Learning (MGNRL) framework for the latent representation of nodes in the network. Game theory and deep learning concepts have been widely used in the education field also. Vos et al. [ 95 ] have combined game theory and deep learning to understand students’ motivation in education. Urbani [ 101 ] has combined game theory and deep learning for music genre classification. Han and Jiequn [ 63 ] have concluded that their approach of Deep Learning Approximation for Stochastic Control Problems should apply to broad areas, including dynamic game theory with more than one agent, dynamic resource allocation with several resources and demands, and properties management with large portfolios. Woo [ 137 ] has used game-theoretic complex analysis for nuclear security by addressing non-zero-sum algorithms. In the research, Woo has used deep learning for data processing, where the neural network is used for wiretapping. Table 2 summarizes critical game-theoretic models/concepts used to model various deep learning/artificial systems.

4 Application of game theory in deep reinforcement learning

This section addresses several contributions to which game models are used to solve problems in reinforcement learning.

Over the past 4–5 years, reinforcement learning has become an exciting area of research [ 136 ]. Reinforcement learning has shown significant applications to develop intelligent opponents in computer games. In such applications, CNN plays a vital role in abstracting important traits from grainy and complex data. Games are modeled for appropriate scenarios to test the reinforcement learning algorithms as they provide an essential overview of how good an algorithm can behave in isolated environments. Real-time strategy games are very complex, and planning for short and long-term strategies is very challenging. The work in [ 4 ] addresses the Deep Real-time strategy game for testing advanced artificial intelligence algorithms. It deals with faster learning. It learns at a rate 50,000 times faster than existing real-time strategy games. It is also helpful for moderately observable state spaces and map complexity. Applying Real-time strategy outperforms micro Real-time strategy, StarCraft II, and ELFon advanced system. It is shown that a Deep Q-Network agent outperforms random-play agents most of the time (more than 70%) using the Deep Real-time strategy. In recent times, machine learning is extensively being used for multi-agent systems. [ 31 ] discusses the related work on deep multi-agent reinforcement learning, decentralized optimization, hierarchical reinforcement learning, and GANs. Multiple learning agents cause training non-stationary and even lead to non-stable training or unexpected results. A ‘Learning with Opponent-Learning Awareness’ (LOLA) approach is explored in the paper where the agents try to learn by opponents’ learning. The LOLA learning approach considers the effect of one agent’s policy on the other agents’ policy. The results demonstrate that the interaction of two LOLA agents goes for tit-for-tat, and after that, cooperation in the iterated prisoners’ dilemma (IPD). LOLA gains higher payoffs compared to a naïve learner, and it is robust against the application of higher-order gradient-based approaches. On the other hand, LOLA agents move towards the Nash equilibrium for infinitely repeated matching pennies games. The LOLA update scheme can be effectively applied using an advanced likelihood ratio policy gradient estimator and build the scheme suitable for model-free reinforcement learning. Focusing on the learning of the opponent, LOLA agents learn to coordinate out of self-interest.

AlphaGo was the first program to win the game Go against a world champion. AlphaGo computed positions and moves in tree search with the help of deep neural networks. These neural networks are trained using supervised learning depending upon the data of moves of humans and by reinforcement learning achieved from self-play. Algorithms are developed based on reinforcement learning without experimental data of humans, supervision, or level of knowledge of the domain apart from game rules. AlphaGo is considered its own teacher. A neural network is trained to predict the selection of moves forAlphaGo and find the games’ winner. This neural network improves tree search by enhancing quality move selection and better self-play in the next iteration [ 120 ]. Tsetlin Automaton is a learning mechanism. A single integer is assumed to be a memory element; the model learns the preferable action in stochastic environments by decrement and increment operations. The Tsetlin is introduced to solve complex pattern identifying problems with simple formulas formed by a set of Tsetlin Automata. The Tsetlin Machine gives satisfying accuracy compared with Logistic Regression, Random Forests, Naive Bayes Classifier, Decision Trees, SVMs, and Neural Networks. The significant features like accuracy, interpretability, and computational complexity make the Tsetlin Machine a prevalent tool [ 38 ]. Reinforcement learning is capable of representing psychological and neuro-scientific features of animal behavior [ 136 ]. To apply reinforcement learning in complex real-world situations, the agents have a difficult task: they need to properly formulate the actual environment obtained from multi-parameters sensory inputs and apply them to relate the past experience to new instances. A deep Q-network is used for training deep neural networks to build an artificial agent. The agent learns rules from multi-parameters sensory inputs by reinforcement learning. The agent has experimented on classic Atari 2600 games. The performance of the model outperforms other existing algorithms [ 85 ]. To achieve intelligence in the shared environment with multiple agents, Lanctot et al. [ 69 ] have combined unified game theory with deep reinforced learning. To mitigate societal challenges in learning, education Fang [ 30 ] has used deep reinforced learning. Motivated by the urgent need in green security domains such as protecting endangered wildlife from poaching and preventing illegal logging, Yu et al. [ 144 ] have proposed game-theoretic and deep reinforced based learning models to optimize patrols conducted by law enforcement agencies. Xiao et al. [ 139 ] have used a game-theoretical deep reinforcement-based learning technique to secure mobile crowd-sensing.

5 Application of game theory in GAN

This section addresses some contributions in which game models are used to solve problems in GAN. Figure 5 shows a generalized architecture of GAN.

figure 5

Generalized Generative Adversarial Network Structure

The contribution in [ 123 ] sets up a connection amidst a deep learning model and a game-theoretic approach. This work introduces the application of deep learning to solve game-theoretic problems. Some techniques have been addressed to speed up deep learning and gradient-based approaches in actions with continuity for multi-agent adversarial games. On the other hand, multiple GANs are developed as robust distributed games. A Bregman-based strategic deep learning algorithm is introduced for finding robust distributed Nash equilibria, and it is also checked upon in image synthesis and picture classification. GANs are modeled as a min-max game, and a fast learning algorithm using Bregman divergence is explored. A comparative study on the performance of the Bregman-based algorithm with six algorithms is also shown. Table 3 shows a mapping of components of a game with the components of the deep learning network. The work in [ 16 ] shows that deep learning is unsafe for changes in data distribution. So, a deep learning network like CNN is dangerous for adversarial scenarios. An adversarial learning model is designed for supervised learning. Adversarial scenarios are modeled by a game-theoretic approach to the conduct of deep learning. A smart antagonist and a deep learning model interact with each other. The interaction is represented as a two-person sequential non-cooperative Stackelberg game, and stochastic payoff functions are formulated.

Nash equilibrium is determined for the Stackelberg game. It finds a pair of strategies (genetic operations and weights of learners) by which either learner or antagonist can achieve no better payoff by deviating strategies. The performance of the algorithm is determined under various strategy combinations on the dataset of MNIST handwritten digits. The Nash equilibrium seems to be robust to adversarial environments. The outcomes show that game theory and stochastic optimization methods can be applied to study susceptibilities in performance for deep learning models.

In unsupervised learning, GANs have become one of the most effective techniques for self-learning the given distribution of a specific dataset [ 36 ]. In [ 41 ], the tragedy is modeled as a two-player zero-sum game between two players: a discriminator D and a generator G. The generator’s objective are to generate fake images from random noise, and the objective of the discriminator is to categorize on fake and real images correctly. The paper addresses applications of generalized zero-sum games or multi-player zero-sum games to formulate and train GANs. The results signify that by applying a multi-player setup, the severity of the mode collapse problem in GANs can be reduced to generate a diverse set of fake images. It is also demonstrated on different fake toy distributions and real-world datasets like MNIST & CIFAR10.

Several real-life scenarios can be represented as large-scale imperfect information games. The Nash equilibria are initially computed to handle such problems. The authors in [ 47 ] address a new, without prior knowledge-based, scalable and end-to-end method for learning approximate Nash equilibria effectively. The technique integrates deep learning with fictitious self-play. While implemented on Leduc poker, NFSP finds a Nash equilibrium; alternatively, conventional reinforcement learning approaches diverge. In Limit Texas Hold’em, a large-scale poker game, NFSP learns a strategy that determines the performance, superhuman algorithms with the help of strong domain expertise.

Today, GAN is among the most popular frameworks in Deep Learning [ 36 ]. The research on applications of game-theoretic techniques in developing deep learning algorithms has received a great response from the research community. Generative Adversarial Network Games (GANGs) are developed, and they are modeled as a finitely long zero-sum game between a discriminator D and a generator G by applying mixed strategies. Resource-bounded best responses (RBBRs) and source bounded Nash Equilibrium (RB-NE) are defined as a pair of mixed strategies. By using these techniques, neither G nor D can achieve a better RBBR. The RB-NE solution is better than the local Nash equilibria. It detects errors of gradient descent-based escaping local optima and applies them to approximate best response computations. GANGs are solved with the parallel Nash Memory technique that slowly converges to an RB-NE to validate this approach. The results show that it outperforms standard GAN setups and demonstrate that it performs well with typical GAN problems [ 97 ].

A mixed Nash equilibrium of GANs is determined in [ 56 ] besides finding a pure optimal strategy. The probability distributions over pure strategy are optimized for the network. This paper also discusses well-known prox methods, which are applied to solve two-player games with finite strategies. It can even be drawn out to continuously many strategies. Therefore, it is suitable for training GANs. A proof is given to demonstrate the convergence rates to learning the mixed Nash equilibrium. It is further shown that heuristic approaches significantly reduce memory and computational costs. As a result, they become simple algorithms having complexity per-iteration almost as cheap as SGD. The experimental results prove out that the performance of the algorithms significantly improves, and they are better than popular techniques such as Adam, SGD, and RMSProp.

GANs are used for generating images and learning in a semi-supervised way [ 19 ]. In two-player games, a single discriminator plays a vital role in detecting fraudulent specimens and determining labels, and it also guesses the data without the labels. A triple generative adversarial net (Triple-GAN) is developed in [ 18 ] to solve the problems. Triple-GAN has three players- a classifier, a generator, and a discriminator. The tasks of the generator and the classifier are to determine the conditional distributions between labels and images. On the other hand, the task of the discriminator is to detect fake image-label pairs. The paper addresses compatible utilities that characterize distributions by the generator and the classifier and converge to the data distribution. The outcomes on different datasets show that Triple-GAN is an improved model that can parallelly perform classification results among deep generative models.

GAN can generate realistic images [ 19 , 36 , 130 ]. The visual presentation of objects is affected by their shape geometry which is crucial information but not considered in existing generative models. The work in [ 67 ] addresses the Geometry-Aware Generative Adversarial Networks (GAGAN) that also accounts for geometric information into the image generation technique. In GAGAN, the generator generates latent variables based on the probability distribution of a statistical shape model. By mapping the outcome of the generator with a coordinate table with the help of a differentiable geometric transformation, the geometry of the objects is studied, and the relation from the priorly generated object is determined. The outcomes on face generation show that the GAGAN can generate realistic pictures of faces with random facial traits of superior nature to existing GAN-based models. This approach also enhances the quality of the generated images.

In [ 129 ], a minimax game model is formulated for discriminative (classifier) training. In this game, a generator is considered a player that tries to generate training samples. A different player, the discriminator, has to classify the specimens generated by the generator correctly. The generator detects the Bayes decision boundary and only chooses training samples from there. Alternatively, the discriminator only performs random guesses. Rigorous research has been going on GAN for the last few years, but it is still very challenging to conclude which algorithm is most effective [ 80 ]. The authors discuss a neutral, multi-faceted large-scale empirical study and also compare the results with existing models. The performances of various models are similar under enough optimization of hyper-parameter and random restarts. Different data sets are used to find the precision and recall to deal with certain limitations. The experimental results show that the GAN research is moving towards more systematic and objective evaluation approaches. The work also highlights that not a single tested algorithm consistently outperforms the other algorithms.

Various training methods and architectural characteristics are developed for GANs. The works in [ 109 ] highlight two utilizations of GANs, which are semi-supervised learning and realistic image generation. The techniques discussed in the paper compares the outcomes in semi-supervised classification on various datasets such as CIFAR-10, MNIST, and SVHN. The quality of spawned images is of very superior quality, and the model generates MNIST samples that are very hard to identify for humans. On the other hand, CIFAR-10 samples produce a human inaccuracy rate of nearly 21.3%. The paper also shows that approaches facilitate the model to learn significant features of ImageNet classes.

GANs exploit supervised learning to estimate a cost function. GANs can use a supervised ratio estimation scheme to estimate multiple cost functions such as KL divergence applied for maximum likelihood estimation. Training GANs is to find Nash equilibria in elevated dimensional, continuous, non-convex games [ 35 ]. CNNs are helpful to identify objects in an image, but it is tough to find the position of an object relative to others. A novel solution to this problem is trusting “Capsules” – a logistic unit that confirms the object’s existence and orientation. This work also highlights incorporating the Capsule into the Discriminator of the GANs, which can yield a better classification loss and faster convergence. The results are shown both qualitatively and quantitatively [ 33 ].

The graph representation learning techniques are divided into a couple of categories: generative models that learn the connectivity relation in the graph and discriminative models that determine the probability of an edge between any two vertices. Graph representation learning aims to store each vertex in a graph into a low-dimensional vector. A new learning framework GraphGAN is introduced in [ 133 ]. GraphGAN represents the generative model and discriminative model as a game-theoretical minimax game. The generative model generates accurate connectivity distribution over all vertices and yields forge samples to confuse the discriminative model for a given vertex.

On the other hand, the discriminative model aims to detect whether the sample vertex is real or caused by the generative model. The generative model is implemented using a novel graph softmax. Extensive investigations on real-world datasets confirm that GraphGAN performs well in different applications, e.g., node classification, link prediction, and recommendation.

Lipschitz regularization theory and algorithmic technique for a new Loss-Sensitive Generative Adversarial Network (LS-GAN) are applied in [ 105 ]. It instructs a loss function to differentiate original and fake specimens by precise margins, and a generator generates realistic samples by reducing their losses. The LS-GAN regularizes its loss function by incorporating Lipschitz regularity condition on the probability of actual data produces a regularized model, which better generates new data from a fair number of training specimens as compared to conventional GAN. A generalized LS-GAN (GLS-GAN) is discussed. Other GAN models are compared with LS-GAN and GLS-GAN. LS-GAN is elaborated for supervised and semi-supervised learning problems. In [ 58 ], a data-driven model is explored, known as generative adversarial privacy (GAP). GAP helps the data holder to study the privacy policy from the data. Determining the optimal privacy policy is modeled as a constrained minimax game between a privatizer and an adversary. GAP offers privacy against robust information-theoretic adversaries. GAP is applied for multi-dimensional Gaussian mixture models and large datasets.

The work in [ 45 ] shows an approach for synthesizing a road network by applying GANs. As per the suggested approach, the road network is first converted into a binary image in which pixel intensities represent the presence or absence of streets. Then, the GAN is trained to automatically synthesize the arbitrarily sized street networks, and the generator tries to generate something similar to the original patch. A graph-based representation is extracted from the generated images. The method shows that it can be applicable to synthesize large-scale street networks. Many research articles have already applied recurrent neural networks for music generation. A recent significant contribution, the WaveNet model, was invented by DeepMind. It shows that CNNs can also be used to generate realistic musical waveforms. It has also experimented with that CNNs can generate melody one bar after another in the symbolic representation. GAN is applied in this scenario in which the generator and Discriminator play their roles. A discriminator learns the distributions of melodies. A new conditional approach is exploited based on available prior knowledge. As a result, the model can produce melodies by following a chord sequence or by conditioning on the melody of the preceding, among other capabilities. The proposed MidiNet can also be applied to generate music with multiple MIDI channels. MidiNet and MelodyRNN models are compared, and the results show that MidiNet performs similarly to MelodyRNN models. MidiNet’s melodies are captivating [ 140 ].

Story Cloze Test is an intelligent machine comprehension tool to understand natural language problems. The idea behind the device is too many story tests within the dataset that develops common sense assuming capability. The training data is almost unsupervised in this problem in which context documents having a single positive sentence that is understood from the situation. To deal with the problem, GANs are applied to generate fake sentences. The work in [ 131 ] also develops Conditional GANs (CGANs) in which the situation influences the generator. The experimental outcomes display the efficacy of the CGANs in the discriminating sentence in practical story reading comprehension tasks. GANs can be modeled by a two-player minimax game [ 36 ]. Durugkar et al. [ 27 ] presents the Generative Multi-Adversarial Network (GMAN), which uses multiple discriminators. GMAN is trained with the original, untampered objective. Numerous design perspectives are addressed in the paper in which the Discriminator plays various roles ranging from formidable adversary to forgiving teacher. It is shown that GMAN generates higher quality samples in a lesser number of iterations compared to GAN.

Another significant work explores a novel training process, SGAN, in which multiple adversarial local pairs of networks are trained independently to train a global supervising pair of networks against them. The objective of this process is to train the international pair with the corresponding opponent for better performances. This method tries to increase the chances to continue learning for the global pair. It also prevents trapping in an unsatisfactory local minimum and oscillations in results. In the SGAN training process, the global generator and Discriminator are trained using the local discriminators and generators, respectively. Alternatively, the local networks are trained with their fixed local opponents. The experimental results based on a small scale and real-world problems show that this method outperforms existing approaches [ 14 ]. Pfau and Vinyals [ 103 ] shows that GANs can be modeled as actor-critic frameworks in a scenario wherein the actor cannot influence the payoff. The strategies are formulated for balancing training for each class of models. GANs and reinforcement learning algorithms are analyzed in-depth with more complicated information flow. The contributions in the paper encourage both GAN and reinforcement learning research communities to develop more efficient algorithms for multilevel optimization with deep networks.

Consider a scenario where learning properties from the behavior of an expert are required without interacting with the expert or accessing a reinforcement framework. A solution is to extract the payoff function of the expert with inverse reinforcement learning that leads to extract a property [ 136 ]. The work in [ 53 ] shows that a specific instantiation of the model relates to imitation learning and GANs from which a model-free imitation learning algorithm is derived. The algorithm outperforms the existing model-free methods in imitating various behaviors in real environments. The Discriminator in GAN works as a classifier with the sigmoid cross-entropy loss function. It is observed that the loss function may struggle from the vanishing gradients problem during the learning process. Least Squares Generative Adversarial Networks (LSGANs) is introduced to solve this problem [ 82 ]. LSGANs embrace the least squares loss function for the Discriminator. It is shown that the minimization of the objective function of LSGAN minimizes the Pearson χ2 divergence. It experimented on LSUN and CIFAR-10 datasets and observed that LSGANs could produce superior quality images than GANs. LSGANs are more stabilized during the learning process. It is concluded from two experiments that LSGANs are more stable than GANs.

GAN’s learning dynamics are not still well understood. To deal with GAN’s learning dynamics, a model is explored to demonstrate various problematic convergence behaviors such as vanishing gradient, mode collapse, diverging, or oscillatory behavior [ 73 ]. The model also determines the first convergence bounds for parametric GAN dynamics. It is also shown that a GAN with an optimal discriminator may converge while a first order approximation of the discriminator steps becomes unstable GAN dynamics and mode collapse. The model finds a particular challenge in GAN while training that is called discriminator collapse. The work in [ 9 ] presents an equilibrium enforcing technique and a loss function by Wasserstein distance for training auto-encoder-based GANs. This technique considers the role of generator and Discriminator while training. It also finds an estimated convergence measure, fast and stable training, and a high-quality sample. It is also controlled the trade-off between sample diversity and quality.

Two information retrieval models: the generative retrieval predicting related documents based on a query and the discriminative retrieval indicating relation based on a query document pair, are addressed in [ 132 ]. A minimax game is modeled to optimize both models iteratively. The discriminative model tries to differentiate samples from labeled and unlabelled data. It helps to train the generative model for fitting the distribution of samples based on the query. Alternatively, the generative model generates fake samples for the discriminative model intending to minimize its discrimination objective. (i) the generative model learns to fit the relational distribution of documents with the help of samples from the discriminative model, (ii) the discriminative model can use the unlabelled data chosen by the generative model to perform a better appraisal for document ranking. The experimental outcomes show the efficacy of the model. GAN training may become prosperous, but the trained distribution may not achieve excellent performance for target distribution. Generalization occurs for a weaker metric, neural network length. It is demonstrated that an approximate pure equilibrium is obtained from the Discriminator or generator game. The existence of equilibrium inspires the MIX+GAN protocol [ 5 ].

The authors in [ 94 ] address a model to deal with the mode collapse problem in GAN. The Kullback-Leibler (KL) and reverse KL divergences are combined into a unified objective function. As a result, it can use various statistical properties from the divergences to properly diversify the approximated density in capturing multi-modes. The contribution in this paper includes dual Discriminators generative adversarial nets (DDGAN) having double discriminators and a single generator. In this model, one of the discriminators pays high scores for samples obtained from data, whereas another discriminator supports data obtained from the generator. Alternatively, the generator yields samples to fool both the discriminators. It is shown that the maximal discriminators optimize the generator of DDGAN, which lessens to minimizing KL and reverse KL divergences between data distribution and the distribution from the data produced by the generator. Extensive experiments are managed on synthetic and real-world large-scale datasets such as MNIST, CIFAR-10, STL-10, ImageNet. The experimental outcomes show the efficacy of the model.

A method is explored in [ 52 ] for training GANs with discrete data that exploits the estimated difference from the Discriminator to determine significant weights for generated samples and supplying a policy gradient for training the generator. The significant weights have a robust relationship to the decision boundary of the Discriminator. The name of this method is aptly given boundary-seeking GANs (BGANs). The efficacy of the model is demonstrated with a discrete image and character-based natural language generation. The boundary-seeking objective is also applicable for continuous data shown on Celeba, Large-scale Scene Understanding bedrooms, and Imagenet without conditioning. GANs are best to converge to a local Nash equilibrium. Such a local Nash equilibrium is different from a Nash equilibrium. The work in [ 98 ] models GANs as finite games with mixed strategies, and the method ensures that every local Nash equilibrium is a Nash equilibrium. The authors also introduce another way that monotonically converges to a resource-bounded Nash equilibrium in which computational resources are increased to achieve better solutions. The model shows its efficiency concerning typical GANs and MGANs.

For GANs, the data distribution is learned by a two-player zero-sum game, played between a generator and a discriminator, and both of them aim to optimize their payoffs [ 19 , 36 ]. GANs find Nash equilibrium in which no player can improve payoff by unilaterally deviating its strategy. GANs shows that the training and model distribution difference is the minimum value at an equilibrium point. The divergence has a vital part in the learning process, and each learning step tries to decrease the divergence. We show that this view is overly restrictive. The authors show that GANs can learn distributions when divergence minimization predicts failure [ 29 ]. It is also demonstrated that gradient penalties are equally helpful when applied in other cases in which the divergence minimization does not predict they would be beneficial. GAN training may be more helpful in modeling Nash equilibria.

Structure Generative Adversarial Network (SGAN) has been explored for semi-supervised conditional generative modeling to learn from a small set of separate instances to extract the semantics of significant interest from others in the latent space. It is shown that SGAN has improved extraction and controllability compared to primary frameworks. The model is advantageous to apply to various problems. It outruns existing models regarding the inception score and visual quality of controllable image generation [ 24 ]. GraphSGAN is proposed for graph-based semi-supervised learning by applying GANs [ 25 ]. The scenario is represented as a competitive game between generator and classifier. In the game, the generator generates samples in density gaps at equilibrium. The loss terms cumulatively determine the expected equilibrium. Experiments performed on three standard datasets show the efficacy of the strategy. GraphSGAN model has been thoroughly analyzed. Generated samples minimize the adjacent nodes in density gaps to find decision boundaries prominent. The GraphSGAN is scalable. The experimental results on the DIEL dataset prove the model’s capabilities to execute well on large graphs.

A holistic active learning model by the query synthesis method using GAN is introduced in [ 149 ].

Table 4 summarizes critical game-theoretic models/concepts used to model various GAN architectures. The algorithm adaptively synthesizes training samples for querying to enhance learning speed. The queries are generated according to the uncertainty principle, but the model performs well with other active learning principles. The results are obtained from a large number of experiments to show the efficacy of the approach. In a few settings, the algorithm outperforms existing pool-based methods. An approach for modeling the decision-making of players through agents is introduced in [ 55 ]. The authors claim that artificial agents have the capabilities to act like a human player’s decision-making processes. Such agents can mimic human decision-making and effectively performing testing tools in the game design process. This claim is proved in a crowdsourced decision-making experiment that shows the decisions of human players in a small-scale dungeon-themed puzzle game. Human choices are correlated to the findings of agents [ 127 ]. A decent amount of research has been performed on generative models in computer vision and image classification. These generative models have been made famous in a framework well known as Generative Adversarial Networks. The work in [ 51 ] presents the original model proposed by Good fellow et al. and extensions of the original model and compares these models.

The theory of mind model is introduced in [ 143 ]. The model discusses the intentions and goals of opponents to optimize mutual interactions. Control theory and game theory are applied to model the game theory of mind. The plans are represented as value functions, also known as utility or rewards. The joint value functions are complex since the behaviors of the players are optimized recursively. The joint functions find how player1 determines the value function of player2. Similarly, player2 determines the value function of player1, and so on. If the degree of recursion is bounded, then the players/opponents need to estimate the degree of recursion of the opponent to respond optimally. The contribution shows that it is possible to deduce whether players infer about each other and quantify their value functions based on selections in sequential games. Then, generative models of strategies with and without inference are compared. The comparison is shown using simulated and real data from a ‘stag-hunt.’ It also shows that the same behavior can be found by optimizing the utility function for altruistic agents.

Multiple decision-makers are usually involved in handling real-world problems. They fix their own goals or destinations. Game theory has certain limitations; its applicability is restricted by the intractability of determining equilibrium in large games. Other challenges are that the parameters in the games are sometimes unknown, and the players in real-world scenarios may not be entirely rational. The work in [ 120 ] addresses how learning with game theory can solve societal challenges such as security and sustainability.

Table 5 outlines statistics on the number of articles of applications of game theory in deep learning, reinforcement learning, and exclusively in GAN. Machine learning, reinforcement learning, and deep learning have achieved tremendous responses in different fields and can be applied to overcome the limitations of the game-theoretic models [ 76 , 84 , 136 ].

Table 6 gives statistics of the scientific works discussed in the paper based on the number of articles published, average citations, and maximum citations of applications of game theory in deep learning, reinforcement learning, and exclusively in GAN.

Figure 6 shows the statistics of Table 6 with the help of a bar chart that will give a clear insight to the readers. Tables 5 and 6 , and Fig. 6 individually show that game theory has more applications in GAN over other deep learning models and reinforcement algorithms. GAN has received tremendous responses from the scientific community.

figure 6

Average citations, maximum citations based on the articles discussed on applications of game theory in deep learning, reinforcement learning, and exclusively in GAN

The datasets used in most of the research articles in this study are as follows: MNIST, CelebA, CIFAR-10, BBBC039, initial dataset, transfer dataset, SVHN, STL-10, ImageNet, DIEL, etc.

6 Implementation environment

The current study mainly focuses on deep reinforcement learning and GAN models. There are various real-time and business applications of GAN and reinforcement learning. The current study addresses several diverse problems. Some significant implementation environments of GAN and deep reinforcement learning models are described in this section based on the existing literature. GAN is primarily used to generate samples/images for various image datasets. GAN can be used to create images of human faces, which are used for different purposes. GANs also help to create realistic images. The cartoon industry is a popular entertainment industry where GAN is used to generate various cartoon characters. Another critical application of GAN is image to image translation. Text to image translation is another vital application of GAN. GAN can be used to generate new human postures which have security, healthcare, and entertainment applications. Another exciting application is generating emojis from typical images frequently used in various social network platforms, mobile applications for entertainment purposes. Photos are edited using GAN models. Photo blending is also done using GAN models. GAN is used for video prediction and 3D object generation. GAN can be used for improving cybersecurity in various places and improving healthcare sectors. It can create artistic skills; it can generate exact or similar images of a painting. It can even cause fake videos. The quality of an image can also be improved using GAN that serves several purposes. De-noising on images is also done using GAN. Alternatively, deep reinforcement learning has excellent application in real-time systems and industries. It is used for the implementation of a self-driving car and industry automation. Deep reinforcement learning is also used in NLP applications. The Healthcare system has also adopted deep reinforcement learning as a tool. It successfully performs various tasks such as automatic disease prediction etc. It has other robust applications in news recommendation and gaming. It also has applications to build automated robots. It has other significant applications such as resource management in computer clusters, traffic light control, web system configuration, advertisement, etc. Reduction in energy consumption, online recommendation systems are some industrial applications.

The performances of the proposed study can be evaluated through Sensitivity, Specificity, recall. Specificity is defined as the proportion of actual negatives, which got predicted as the negative. Sensitivity is a measure of the balance of real positive cases that got expected as positive. The recall is the measure of our model correctly identifying True Positives [ 87 ].

7 Future research direction and challenges

The previous sections of the present work discuss the linkages and applications of game theory in deep learning. The current section aims to discuss future research challenges related to the application of game theory in deep learning. To narrate and list out all possible future challenges is a complex task as different literature uses different simulation environments, simulation tools, different data sets, and different experimental conditions. Nevertheless, the current section discusses some of the critical representative future research directions. The future research direction is described in Table 7 below. Table 7 also provides hints, literary way, and association of game theory and deep learning to get the breakthrough in solving future research challenges.

In conclusion, researchers and scientists have to do extensive developmental research and efforts in the direction described above to overcome the problems and challenges faced by the future of deep learning, such as genomics and medical imaging. Besides, more techniques and further inspiration are required to develop new deep learning approaches. New methodologies for complex and challenging problems would be necessary as collaborative efforts by various research communities need to be carefully addressed.

8 Conclusions

Games play a vital role in the advancement of artificial intelligence. Games are frequently used as training mechanisms in learning algorithms, such as reinforcement learning and imitation learning. The paper highlights that AI-enabled and deep learning-enabled multi-agent systems are also developed to compete or cooperate to complete a goal with the help of game theory. In real-time scenarios, deep learning frameworks need to face situations with imperfect information. Paper considers all 31 articles in which game theory and deep learning coincide described in Table 1 . It is discussed in the paper that DeepMind’sAlphaGo works with the help of partial knowledge to strategically make superior to the world’s most efficient human in the game of Go.

The paper also puts ample light on the application of game theory to construct adversarial networks. An essential property of an adversarial network is that a closed-form loss function is not needed. Some networks have the capability of finding their loss function. A significant disadvantage of adversarial networks is difficulties in training them. Adversarial learning models find a Nash equilibrium for a two-player non-cooperative game. The paper addresses applications of game theory in artificial neural networks, reinforcement learning, and deep learning. The main focus is to focus on the construction of GAN by applying different game models. The paper’s contribution may help the researchers in game theory, deep learning, and artificial intelligence to acquire a large number of ideas about inter-domain research and contribute holistic works in these domains. The paper also addresses various challenges and future directions of the identified area that will help researchers explore. In the future, behavioral game theory can be applied to model deep learning networks and mimic human neural networks. The combination of the research areas will open several paths for the researchers.

Abraham I, Dolev D, Gonen R, Halpern J (2006) Distributed computing meets game theory: robust mechanisms for rational secret sharing and multiparty computation. In: Proceedings of the twenty-fifth annual ACM symposium on principles of distributed computing, pp 53–62

Agrawal A, Jaiswal D (1981) When machine learning meets ai and game theory

Alafif T, Tehame AM, Bajaba S, Barnawi A, Zia S (2021) Machine and deep learning towards COVID-19 diagnosis and treatment: survey, challenges, and future directions. Int J Environ Res Public Health 18(3):1117

Article   Google Scholar  

Andersen PA, Goodwin M, Granmo OC (2018) Deep RTS: a game environment for deep reinforcement learning in real-time strategy games. In: 2018 IEEE conference on computational intelligence and games (CIG). IEEE, pp 1–8

Google Scholar  

Arora S, Ge R, Liang Y, Ma T, Zhang Y (2017) Generalization and equilibrium in generative adversarial nets (gans). In: Proceedings of the 34th international conference on machine learning-Volume 70. JMLR.org, pp 224–232

Aslan S, Vascon S, Pelillo M (2020) Two sides of the same coin: improved ancient coin classification using graph transduction games. Pattern Recogn Lett 131:158–165

Balliet D, Mulder LB, Van Lange PA (2011) Reward, punishment, and cooperation: a meta-analysis. Psychol Bull 137(4):594–615

Baum EB (1988) On the capabilities of multilayer perceptrons. J Complex 4(3):193–215

Article   MathSciNet   MATH   Google Scholar  

Berthelot D, Schumm T, Metz L (2017) Began: boundary equilibrium generative adversarial networks. arXiv preprint arXiv:1703.10717

Brams SJ (2003) Negotiation games: applying game theory to bargaining and arbitration, vol 2, Psychology Press

Caicedo JC, Roth J, Goodman A, Becker T, Karhohs KW, Broisin M, … Carpenter AE (2019) Evaluation of deep learning strategies for nucleus segmentation in fluorescence images. Cytometry Part A 95(9):952–965

Carse J (2011) Finite and infinite games. Simon and Schuster

Carter E (2019) Deep learning for robust meta-analytic estimation

Chavdarova T, Fleuret F (2018) Sgan: an alternative training of generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9407–9415

Cheng J, Dong L, Lapata M (2016) Long short-term memory-networks for machine reading. arXiv preprint arXiv:1601.06733

Chivukula AS, Liu W (2017) Adversarial learning games with deep learning models. In: 2017 international joint conference on neural networks (IJCNN). IEEE, pp 2758–2767

Chivukula AS, Liu W (2018) Adversarial deep learning models with multiple adversaries. IEEE Trans Knowl Data Eng 31(6):1066–1079

Chongxuan LI, Xu T, Zhu J, Zhang B (2017) Triple generative adversarial nets. In: Advances in neural information processing systems, pp 4088–4098

Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath AA (2018) Generative adversarial networks: an overview. IEEE Signal Process Mag 35(1):53–65

Darwish A, Hassanien AE, Das S (2019) A survey of swarm and evolutionary computing approaches for deep learning. Artificial intelligence review, pp 1–46

Dasgupta P, Collins JB (2019) A survey of game theoretic approaches for adversarial machine learning in cybersecurity tasks. arXiv preprint arXiv:1912.02258

David B (2016) Grammars for games: a gradient-based, game-theoretic framework for optimization in deep learning. Front Robot AI 2:39

Davis M, Maschler M (1965) The kernel of a cooperative game. Naval Res Logist Q 12(3):223–259

Deng Z, Zhang H, Liang X, Yang L, Xu S, Zhu J, Xing EP (2017) Structured generative adversarial networks. In: Advances in neural information processing systems, pp 3899–3909

Ding M, Tang J, Zhang J (2018) Semi-supervised learning on graphs with generative adversarial nets. In: Proceedings of the 27th ACM international conference on information and knowledge management. ACM, pp 913–922

Duchi J, Hazan E, Singer Y (2011) Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res 12(Jul):2121–2159

MathSciNet   MATH   Google Scholar  

Durugkar I, Gemp I, Mahadevan S (2016) Generative multi-adversarial networks. arXiv preprint arXiv:1611.01673

Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, … Dean J (2019) A guide to deep learning in healthcare. Nat Med 25(1):24–29

Fedus W, Rosca M, Lakshminarayanan B, Dai AM, Mohamed S, Goodfellow I (2017) Many paths to equilibrium: GANs do not need to decrease a divergence at every step. arXiv preprint arXiv:1710.08446

Fei F (2019) Integrate learning with game theory for societal challenges. Proceedings of the 28th International Joint Conference on Artificial Intelligence. AAAI Press

Foerster J, Chen RY, Al-Shedivat M, Whiteson S, Abbeel P, Mordatch I (2018) Learning with opponent-learning awareness. In: Proceedings of the 17th international conference on autonomous agents and MultiAgent systems. International Foundation for Autonomous Agents and Multiagent Systems, pp 122–130

Fukushima K (1980) Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern 36(4):193–202

Article   MATH   Google Scholar  

Gadirov H (2018) Capsule architecture as a discriminator in generative adversarial networks

Glorot X, Bordes A, Bengio, Y. (2011) Deep sparse rectifier neural networks. In: Proceedings of the fourteenth international conference on artificial intelligence and statistics, pp 315–323

Goodfellow I (2016) NIPS 2016 tutorial: generative adversarial networks. arXiv preprint arXiv:1701.00160

Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672–2680

Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT press

MATH   Google Scholar  

Granmo OC (2018) The Tsetlin machine-a game theoretic bandit driven approach to optimal pattern recognition with propositional logic. arXiv preprint arXiv:1804.01508

Graves A, Mohamed AR, Hinton G (2013) Speech recognition with deep recurrent neural networks. In: 2013 IEEE international conference on acoustics, speech and signal processing. IEEE, pp 6645–6649

Guo X, Singh S, Lewis R, Lee H (2016) Deep learning for reward design to improve Monte Carlo tree search in atari games. arXiv preprint arXiv:1604.07095

Gupta S (2018) Multi-player generative adversarial networks. In: 2018 international high performance extreme computing conference (HPEC). IEEE

Gurram P, Kwon H (2014) Coalition game theory based feature subset selection for hyperspectral image classification. In: 2014 IEEE geoscience and remote sensing symposium. IEEE, pp 3446–3449

Halpern JY (2008) Computer science and game theory. The New Palgrave Dictionary of Economics 1–8:984–994

Hang S, Liu Q, Xia S (2018) Multi-granularity network representation learning based on game theory. 2018 IEEE international conference on data mining workshops (ICDMW). IEEE, p 2018

Hartmann S, Weinmann M, Wessel R, Klein R (2017) StreetGAN: towards road network synthesis with generative adversarial networks

He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

Heinrich J, Silver D (2016) Deep reinforcement learning from self-play in imperfect-information games. arXiv preprint arXiv:1603.01121

Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554

Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv preprint arXiv:1207.0580

Hinton G, Deng L, Yu D, Dahl GE, Mohamed AR, Jaitly N, Senior A, Vanhoucke V, Nguyen P, Sainath TN, Kingsbury B (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process Mag 29(6):82–97

Hitawala S (2018) Comparative study on generative adversarial networks. arXiv preprint arXiv:1801.04271

Hjelm RD, Jacob AP, Che T, Trischler A, Cho K, Bengio Y (2017) Boundary-seeking generative adversarial networks. arXiv preprint arXiv:1702.08431

Ho J, Ermon S (2016) Generative adversarial imitation learning. In: Advances in neural information processing systems, pp 4565–4573

Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

Holmgård C, Liapis A, Togelius J, Yannakakis GN (2014) Generative agents for player decision modeling in games. In: FDG

Hsieh YP, Liu C, Cevher V (2018) Finding mixed Nash equilibria of generative adversarial networks. arXiv preprint arXiv:1811.02002

Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708

Huang, C., Kairouz, P., Chen, X., Sankar, L. and Rajagopal, R., 2018. Generative adversarial privacy. arXiv preprint arXiv:1807.05306.

Hubel DH, Wiesel TN (1959) Receptive fields of single neurones in the cat's striate cortex. J Physiol 148(3):574–591

Hubel DH, Wiesel TN (1968) Receptive fields and functional architecture of monkey striate cortex. J Physiol 195(1):215–243

Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167

Jain AK, Mao J, Mohiuddin KM (1996) Artificial neural networks: a tutorial. Computer 29(3):31–44

Jiequn H (2016) Deep learning approximation for stochastic control problems. arXiv preprint arXiv:1611.07422

Johnson ND, Mislin AA (2011) Trust games: a meta-analysis. J Econ Psychol 32(5):865–889

Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980

Kohonen T (1997) Exploration of very large databases by self-organizing maps. In: Proceedings of international conference on neural networks (icnn'97), vol 1). IEEE, pp PL1–PL6

Kossaifi J, Tran L, Panagakis Y, Pantic M (2018) Gagan: Geometry-aware generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 878–887

Lamba, A., 2013. Enhancing awareness of cyber-security and cloud computing using principles of game theory. Int J Adv Manag Technol Eng Sci, 3.

Lanctot M et al (2017) A unified game-theoretic approach to multiagent reinforcement learning. Advances in neural information processing systems

Larichev OI, Moshkovich HM (2013) Verbal decision analysis for unstructured problems , vol 17. Springer Science & Business Media, p 51

Book   MATH   Google Scholar  

Leckie C, Peyam P, Jack R (2018) Deep Learning Based Game-Theoretical Approach to Evade Jamming Attacks. Decision and Game Theory for Security: 9th International Conference, GameSec 2018, Seattle, WA, USA, October 29–31, 2018, proceedings, vol 11199. Springer

LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436

Li J, Madry A, Peebles J, Schmidt L (2017) Towards understanding the dynamics of generative adversarial networks. arXiv preprint arXiv:1706.09884

Li X, Dvornek NC, Zhou Y, Zhuang J, Ventola P, Duncan JS (2019) Efficient interpretation of deep learning models using graph structure and cooperative game theory: application to ASD biomarker discovery. In International conference on information processing in medical imaging. Springer, Cham, pp 718–730

Liang X, Xiao Y (2012) Game theory for network security. IEEE Commun Surv Tutor 15(1):472–486

Article   MathSciNet   Google Scholar  

Lippi M (2015) Statistical relational learning for game theory. IEEE Trans Comput Intell AI Games 8(4):412–425

Liu J, Snodgrass S, Khalifa A, Risi S, Yannakakis GN, Togelius J (2021) Deep learning for procedural content generation. Neural Comput & Applic 33(1):19–37

Lu Y, Kai Y (2020) Algorithms in multi-agent systems: a holistic perspective from reinforcement learning and game theory. arXiv preprint arXiv:2001.06487

Luce RD, Raiffa H (1989) Games and decisions: introduction and critical survey. Courier Corporation

Lucic M, Kurach K, Michalski M, Gelly S, Bousquet O (2018) Are gans created equal? A large-scale study. In: advances in neural information processing systems, pp 700–709

Malherbe C, Umarova RM, Zavaglia M, Kaller CP, Beume L, Thomalla G, Weiller C, Hilgetag CC (2018) Neural correlates of visuospatial bias in patients with left hemisphere stroke: a causal functional contribution analysis based on game theory. Neuropsychologia 115:142–153

Mao X, Li Q, Xie H, Lau RY, Wang Z, Paul Smolley S (2017) Least squares generative adversarial networks. In Proceedings of the IEEE international conference on computer vision, pp 2794–2802

Marchiori D, Nagel R, Schmidt J (2019) Heuristics, thinking about others, and strategic management: insights from behavioral game theory

Marsland S (2014) Machine learning: an algorithmic perspective. Chapman and Hall/CRC

Book   Google Scholar  

Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M., Fidjeland, A.K., Ostrovski, G. and Petersen, S., 2015. Human-level control through deep reinforcement learning. Nature, 518(7540), p.529.

Moran N, Pollack J (2018) Coevolutionary neural population models. In artificial life, pp 39–46

Murugan R, Goel T (2021) E-DiCoNet: extreme learning machine based classifier for diagnosis of COVID-19 using deep convolutional network. J Ambient Intell Humaniz Comput:1–12

Mycielski J (1992) Games with perfect information. Handbook of game theory with economic applications 1:41–70

Myerson RB (1991) Game theory: analysis of conflict. Harvard university press

Myerson RB (2013) Game theory. Harvard university press

Narahari Y (2014) Game theory and mechanism design, vol 4. World Scientific

Nash JF (1950) Equilibrium points in n-person games. Proc Natl Acad Sci 36(1):48–49

Nash J (1951) Non-cooperative games. Ann Math 54:286–295

Nguyen T, Le T, Vu H, Phung D (2017) Dual discriminator generative adversarial nets. In: Advances in neural information processing systems, pp 2670–2680

Nienke V, Van H, Denessen E (2011) Effects of constructing versus playing an educational game on student motivation and deep learning strategy use. Comput Educ 56(1):127–137

Nisan N, Roughgarden T, Tardos E, Vazirani VV (eds) (2007) Algorithmic game theory. Cambridge university press

Oliehoek FA, Savani R, Gallego-Posada J, Van der Pol E, De Jong ED, Groß R (2017) GANGs: generative adversarial network games. arXiv preprint arXiv:1712.00679

Oliehoek FA, Savani R, Gallego J, van der Pol E, Groß R (2018) Beyond local nashequilibria for adversarial networks. arXiv preprint arXiv:1806.07268

Osborne MJ (2004) An introduction to game theory, vol 3, no 3. Oxford university press, New York

Osborne MJ, Rubinstein A (1994) A course in game theory. MIT press

Paola U (2018) Combining deep learning and game theory for music genre classification. BS thesis. UniversitàCa'FoscariVenezia

Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, … Desmaison A (2019) PyTorch: an imperative style, high-performance deep learning library. In Advances in neural information processing systems, pp 8024–8035

Pfau D, Vinyals O (2016) Connecting generative adversarial networks and actor-critic methods. arXiv preprint arXiv:1610.01945

Pinto L, James D, Gupta A (2017) Supervision via competition: robot adversaries for learning tasks. 2017 IEEE international conference on robotics and automation (ICRA). IEEE, p 2017

Qi GJ (2017) Loss-sensitive generative adversarial networks on lipschitz densities. arXiv preprint arXiv:1701.06264

Rabinowitz NC, Perbet F, Song HF, Zhang C., Eslami SM, Botvinick M (2018) Machine theory of mind. arXiv preprint arXiv:1802.07740

Ren K, Zheng T, Qin Z, Liu X (2020) Adversarial attacks and defenses in deep learning. Engineering. 6:346–360

Rudrapal D, Boro S, Srivastava J, Singh S (2020) A deep learning approach to predict football match result. In: Computational intelligence in data mining. Springer, Singapore, pp 93–99

Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Chen X (2016) Improved techniques for training gans. In: Advances in neural information processing systems, pp 2234–2242

Saxe AM, Bansal Y, Dapello J, Advani M, Kolchinsky A, Tracey BD, Cox DD (2019) On the information bottleneck theory of deep learning. J Stat Mech Theory Exp 2019(12):124020

Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45(11):2673–2681

Schuster A, Yamaguchi Y (2010) Application of game theory to neuronal networks. Advances in Artificial Intelligence, 2010

Schuurmans D, Zinkevich MA (2016) Deep learning games. In: Advances in neural information processing systems, pp 1678–1686

Schwenker F, Kestler HA, Palm G (2001) Three learning phases for radial-basis-function networks. Neural Netw 14(4–5):439–458

Sengupta N (2021) Information generation in interactions: the link between evolutionary game theory and evolutionary economics

Shapley LS (1953) A value for n-person games. Contributions to the Theory of Games, 2(28), pp 307–317

Shapley LS (1953) Stochastic games. Proc Natl Acad Sci 39(10):1095–1100

Shermila R, Maghsudi S, Hossain E (2017) Mobile edge computation offloading using game theory and reinforcement learning. arXiv preprint arXiv:1711.09012

Shin M, Joongheon K, Marco L (2019) Auction-based charging scheduling with deep learning framework for multi-drone networks. IEEE Trans Veh Technol 68(5):4235–4248

Silver D, Schrittwieser J, Simonyan K, Antonoglou I, Huang A, Guez A, Hubert T, Baker L, Lai M, Bolton A, Chen Y (2017) Mastering the game of go without human knowledge. Nature 550(7676):354

Srivastava RK, Greff K, Schmidhuber J (2015) Highway networks. arXiv preprint arXiv:1505.00387

Stier J, Gianini G, Granitzer M, Ziegler K (2018) Analysing neural network topologies: a game theoretic approach. Procedia Computer Science 126:234–243

Tembine H (2019) Deep learning meets game theory: Bregman-based algorithms for interactive deep generative adversarial networks. IEEE Trans Cybern

Tennenholtz M (2002) Game theory and artificial intelligence. In Foundations and applications of multi-agent systems. Springer, Berlin, pp 49–58

Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol PA (2010) Stacked denoisingautoencoders: learning useful representations in a deep network with a local denoising criterion. J Mach Learn Res 11(Dec):3371–3408

Von Neumann J (1959) On the theory of games of strategy. Contributions to the Theory of Games, 4, pp 13–42

Wakatsuki M, Fujimura M, Nishino T (2020) A decision making method based on society of mind theory in multi-player imperfect information games. In: Deep learning and neural networks: concepts, methodologies, tools, and applications. IGI Global, pp 317–329

Wang B (2018) From deep learning to deep deducing: automatically tracking down Nash equilibrium through autonomous neural agent, a possible missing step toward gneral AI

Wang Y Discriminative adversarial learning: a general framework for interactive learning

Wang K, Gou C, Duan Y, Lin Y, Zheng X, Wang FY (2017) Generative adversarial networks: introduction and outlook. IEEE/CAA J Autom Sin 4(4):588–598

Wang B, Liu K, Zhao J (2017) Conditional generative adversarial networks for commonsense machine comprehension. In: IJCAI, pp 4123–4129

Wang J, Yu L, Zhang W, Gong Y, Xu Y, Wang B, Zhang P, Zhang D (2017) Irgan: a minimax game for unifying generative and discriminative information retrieval models. In: Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval. ACM, pp 515–524

Wang H, Wang J, Wang J, Zhao M, Zhang W, Zhang F, Xie X, Guo M (2018) Graphgan: graph representation learning with generative adversarial nets. In: Thirty-Second AAAI Conference on Artificial Intelligence.

Wang RQ et al (2019) Scene recognition based on DNN and game theory with its applications in human-robot interaction. arXiv preprint arXiv:1912.01293

Wang EK, Chen C-M, Yiu SM, Hassan MM, Alrubaian M, Fortino G (2020) Incentive evolutionary game model for opportunistic social networks. Futur Gener Comput Syst 102:14–29

Wiering M, Van Otterlo M (2012) Reinforcement learning. Adapt Learn Optim 12:3

Woo TH (2019) Game theory based complex analysis for nuclear security using non-zero sum algorithm. Ann Nucl Energy 125:12–17

Wu TY, Lee WT, Guizani N, Wang TM (2014) Incentive mechanism for P2P file sharing based on social network and game theory. J Netw Comput Appl 41:47–55

Xiao L et al (2017) A secure mobile crowdsensing game with deep reinforcement learning. IEEE Trans Inf Forensics Secur 13(1):35–47

Yang LC, Chou SY, Yang YH (2017) MidiNet: a convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847

Yegnanarayana B (2009) Artificial neural networks. PHI Learning Pvt. Ltd

Yoon JH, Lee BK, Kim BW (2021) A study on GAN algorithm for restoration of cultural property (pagoda). J Korea Soc Comput Inf 26(1):77–84

Yoshida W, Dolan RJ, Friston KJ (2008) Game theory of mind. PLoS Comput Biol 4(12):e1000254

Yu L et al (2018) Deep reinforcement learning for green security game with online information. Workshops at the Thirty-Second AAAI Conference on Artificial Intelligence

Zhang L, Wang W, Li S, Pan G (2019) Monte Carlo neural fictitious self-play: approach to approximate Nash equilibrium of imperfect-information games, arXiv:1903.09569v2

Zheng S, Jayasumana S, Romera-Paredes B, Vineet V, Su Z, Du D, Huang C, Torr PH (2015) Conditional random fields as recurrent neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 1529–1537

Zhou C, Paffenroth RC (2017) Anomaly detection with robust deep autoencoders. In: Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining, pp 665–674

Zhou Y, Kantarcioglu M, Xi B (2019) A survey of game theoretic approach for adversarial machine learning. Wiley Interdiscip Rev Data Min Knowl Discov 9 (3):e1259

Zhu JJ, Bento J (2017) Generative adversarial active learning. arXiv preprint arXiv:1702.07956

Zou J, Huss M, Abid A, Mohammadi P, Torkamani A, Telenti A (2019) A primer on deep learning in genomics. Nat Genet 51(1):12–18

Zurada JM (1992) Introduction to artificial neural systems, vol 8. West publishing company, St. Paul

Download references

Author information

Authors and affiliations.

Department of Computer Science and Engineering, Indian Institute of Information Technology Pune, Pune, Maharashtra, India

Tanmoy Hazra

Institute of Rural Management Anand (IRMA), Post Box No. 60, Anand, Gujarat, 388001, India

Kushal Anjaria

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Tanmoy Hazra .

Additional information

Publisher’s note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Hazra, T., Anjaria, K. Applications of game theory in deep learning: a survey. Multimed Tools Appl 81 , 8963–8994 (2022). https://doi.org/10.1007/s11042-022-12153-2

Download citation

Received : 09 June 2021

Revised : 29 August 2021

Accepted : 03 January 2022

Published : 09 February 2022

Issue Date : March 2022

DOI : https://doi.org/10.1007/s11042-022-12153-2

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Artificial neural network
  • Deep learning
  • Game theory
  • Artificial intelligence
  • Reinforcement learning
  • Find a journal
  • Publish with us
  • Track your research

SAP is an open access publisher of journals covering a wide range of academic disciplines.

Scientific & Academic Publishing

  • Conferences
  • Manuscript Submission

Journal of Game Theory

The Journal of Game Theory publishes original research, survey papers and short notes in game theory and its applications providing rigorous analysis and significant contributions from a methodological, conceptual or mathematical point of view. It welcomes game-oriented papers from all fields of study, including: biology; computer science, economics, engineering; management science, operations research, political science, psychology.

  • Aims & Scope
  • Editorial Board

Special Issues

  • Special Issue Guidelines
  • Special Issue-Call for Papers
  • Published Special Issues

Journal Information

  • Current Issue
  • Most Downloaded Articles
  • Journal Subscription
  • Article Processing Charges
  • Publication Ethics and Publication Malpractice Statement
  • Plagiarism Policy

Captcha Page

We apologize for the inconvenience...

To ensure we keep this website safe, please can you confirm you are a human by ticking the box below.

If you are unable to complete the above request please contact us using the below link, providing a screenshot of your experience.

https://ioppublishing.org/contacts/

IEEE Account

  • Change Username/Password
  • Update Address

Purchase Details

  • Payment Options
  • Order History
  • View Purchased Documents

Profile Information

  • Communications Preferences
  • Profession and Education
  • Technical Interests
  • US & Canada: +1 800 678 4333
  • Worldwide: +1 732 981 0060
  • Contact & Support
  • About IEEE Xplore
  • Accessibility
  • Terms of Use
  • Nondiscrimination Policy
  • Privacy & Opting Out of Cookies

A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.

SAP is an open access publisher of journals covering a wide range of academic disciplines.

Scientific & Academic Publishing

  • Conferences
  • Manuscript Submission

Journal of Game Theory

The Journal of Game Theory publishes original research, survey papers and short notes in game theory and its applications providing rigorous analysis and significant contributions from a methodological, conceptual or mathematical point of view. It welcomes game-oriented papers from all fields of study, including: biology; computer science, economics, engineering; management science, operations research, political science, psychology.

  • Aims & Scope
  • Editorial Board

Yeremia Maroutian

Journal of Game Theory , 2024 13(3), pp. 45-51 Published Online: July 31, 2024

10.5923/j.jgt.20241303.01

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License .

Full-Text HTML Download Full-Text PDF 101 Views   69 Downloads

Rashid Faridnia

Journal of Game Theory , 2024 13(2), pp. 37-43 Published Online: July 27, 2024

10.5923/j.jgt.20241302.03

Full-Text HTML Download Full-Text PDF 889 Views   167 Downloads

Ezio Marchi

Journal of Game Theory , 2024 13(2), pp. 32-36 Published Online: July 27, 2024

10.5923/j.jgt.20241302.02

Full-Text HTML Download Full-Text PDF 125 Views   64 Downloads

Maurício Assuero Lima de Freitas, Helena Cristina Soares Lourenço

Journal of Game Theory , 2024 13(2), pp. 21-31 Published Online: March 28, 2024

10.5923/j.jgt.20241302.01

Full-Text HTML Download Full-Text PDF 770 Views   388 Downloads

Journal of Game Theory , 2024 13(1), pp. 15-20 Published Online: March 04, 2024

10.5923/j.jgt.20241301.03

Full-Text HTML Download Full-Text PDF 271 Views   186 Downloads

Maan T. Alabdullah, Esam A. El – Siedy, Eliwa M. Roushdy, Muner M. Abou Hasan

Journal of Game Theory , 2024 13(1), pp. 7-14 Published Online: January 27, 2024

10.5923/j.jgt.20241301.02

Full-Text HTML Download Full-Text PDF 392 Views   255 Downloads

Shahd H. Alkharaz, Esam A. El – Siedy, Eliwa M. Roushdy, Muner M. Abou Hasan

Journal of Game Theory , 2024 13(1), pp. 1-6 Published Online: January 23, 2024

10.5923/j.jgt.20241301.01

Full-Text HTML Download Full-Text PDF 627 Views   406 Downloads

Ben Dhaou Bourheneddine, Ziad Abderrahmane

Journal of Game Theory , 2023 12(1), pp. 1-11 Published Online: March 02, 2023

10.5923/j.jgt.20231201.01

Full-Text HTML Download Full-Text PDF 948 Views   582 Downloads

Beidi Qiang

Journal of Game Theory , 2022 11(1), pp. 1-5 Published Online: December 23, 2022

10.5923/j.jgt.20221101.01

Full-Text HTML Download Full-Text PDF 1224 Views   654 Downloads

Olivier Lefebvre

Journal of Game Theory , 2021 10(1), pp. 1-9 Published Online: April 30, 2021

10.5923/j.jgt.20211001.01

Full-Text HTML Download Full-Text PDF 2142 Views   1230 Downloads

Issah Musah, Douglas Kwasi Boah, Baba Seidu

Journal of Game Theory , 2020 9(2), pp. 25-31 Published Online: October 28, 2020

10.5923/j.jgt.20200902.01

Full-Text HTML Download Full-Text PDF 12967 Views   2844 Downloads

Subarna Roy

Journal of Game Theory , 2020 9(1), pp. 13-23 Published Online: September 28, 2020

10.5923/j.jgt.20200901.03

Full-Text HTML Download Full-Text PDF 2532 Views   1390 Downloads

Jason J. Wang

Journal of Game Theory , 2020 9(1), pp. 8-12

10.5923/j.jgt.20200901.02

Full-Text HTML Download Full-Text PDF 6529 Views   2613 Downloads

Shahd H. Alkaraz, Essam El-Seidy, Neveen S. Morcos

Journal of Game Theory , 2020 9(1), pp. 1-7

10.5923/j.jgt.20200901.01

Full-Text HTML Download Full-Text PDF 28089 Views   4031 Downloads

Hassan El Kady, Essam El-Seidy

Journal of Game Theory , 2019 8(2), pp. 38-41

10.5923/j.jgt.20190802.03

Full-Text HTML Download Full-Text PDF 2529 Views   1585 Downloads

Wesam B. Zorik, Essam EL-Seidy, Entisarat M. Elshobaky

Journal of Game Theory , 2019 8(2), pp. 32-37

10.5923/j.jgt.20190802.02

Full-Text HTML Download Full-Text PDF 4540 Views   2698 Downloads

Essam El_Seidy, Hassan El Kady, Dieaa I. Nassr

Journal of Game Theory , 2019 8(2), pp. 25-31

10.5923/j.jgt.20190802.01

Full-Text HTML Download Full-Text PDF 5014 Views   2612 Downloads

Journal of Game Theory , 2019 8(1), pp. 16-24

10.5923/j.jgt.20190801.03

Full-Text HTML Download Full-Text PDF 2920 Views   1717 Downloads

Zhao Rui-juan, Zhou Jian-heng

Journal of Game Theory , 2019 8(1), pp. 9-15

10.5923/j.jgt.20190801.02

Full-Text HTML Download Full-Text PDF 3246 Views   2011 Downloads

Alexander Munson

Journal of Game Theory , 2019 8(1), pp. 1-8

10.5923/j.jgt.20190801.01

Full-Text HTML Download Full-Text PDF 2930 Views   1870 Downloads

Special Issues

  • Special Issue Guidelines
  • Special Issue-Call for Papers
  • Published Special Issues

Journal Information

  • Current Issue
  • Most Downloaded Articles
  • Journal Subscription
  • Article Processing Charges
  • Publication Ethics and Publication Malpractice Statement
  • Plagiarism Policy

A Brief Introduction to the Basics of Game Theory

21 Pages Posted: 5 Dec 2011 Last revised: 17 Nov 2017

Matthew O. Jackson

Stanford University - Department of Economics; Santa Fe Institute

Date Written: December 5, 2011

I provide a (very) brief introduction to game theory. I have developed these notes to provide quick access to some of the basics of game theory; mainly as an aid for students in courses in which I assumed familiarity with game theory but did not require it as a prerequisite.

Keywords: Game Theory

JEL Classification: C72

Suggested Citation: Suggested Citation

Matthew O. Jackson (Contact Author)

Stanford university - department of economics ( email ).

Landau Economics Building 579 Serra Mall Stanford, CA 94305-6072 United States 1-650-723-3544 (Phone)

HOME PAGE: http://www.stanford.edu/~jacksonm

Santa Fe Institute

1399 Hyde Park Road Santa Fe, NM 87501 United States

Do you have a job opening that you would like to promote on SSRN?

Paper statistics, related ejournals, social sciences education ejournal.

Subscribe to this fee journal for more curated articles on this topic

Behavioral & Experimental Finance (Editor's Choice) eJournal

Subscribe to this free journal for more curated articles on this topic

Finance Education eJournal

Economics educator: courses, cases & teaching ejournal.

Game Theory Research Paper Topics

Academic Writing Service

This comprehensive guide to game theory research paper topics is designed to assist students and researchers in the field of economics. Selecting a compelling and relevant research topic is a crucial step in the academic journey, and this guide aims to facilitate this process. We provide an extensive list of topics, divided into ten categories, each with ten unique ideas. Additionally, we offer expert advice on how to select a topic from this multitude and how to write a research paper in game theory. Lastly, we introduce iResearchNet’s professional writing services, tailored to support your academic journey and ensure success in your research endeavors.

100 Game Theory Research Paper Topics

Choosing a research paper topic is a critical step in the research process. The topic you select will guide your study and influence the complexity and relevance of your work. In the field of game theory, there are numerous intriguing topics that can be explored. To assist you in this process, we have compiled a comprehensive list of game theory research paper topics. These topics are divided into ten categories, each offering a different perspective on game theory.

Academic Writing, Editing, Proofreading, And Problem Solving Services

Get 10% off with 24start discount code.

  • The role of Nash Equilibrium in game theory
  • The concept of Dominant Strategy and its applications
  • The Prisoner’s Dilemma and its implications
  • The concept of Subgame Perfect Equilibrium
  • Zero-sum games and their significance
  • The role of Mixed Strategy in game theory
  • The concept of Pareto Efficiency in game theory
  • Evolutionary game theory and its applications
  • The concept of Bayesian games
  • The role of Repeated Games in game theory
  • The role of game theory in strategic decision making
  • The impact of game theory on negotiation strategies
  • Game theory and decision making under uncertainty
  • The role of game theory in conflict resolution
  • The impact of game theory on cooperative decision making
  • Game theory and decision making in competitive environments
  • The role of game theory in risk management
  • Game theory and decision making in auctions
  • The impact of game theory on voting strategies
  • Game theory and decision making in social dilemmas
  • The use of game theory in understanding market competition
  • The impact of game theory on economic policy making
  • The role of game theory in understanding oligopolies
  • Game theory and its role in understanding bargaining situations
  • The influence of game theory on economic behavior
  • The role of game theory in understanding public goods and free-riding
  • The impact of game theory on understanding market failures
  • Game theory and its role in understanding strategic trade policy
  • The influence of game theory on understanding labor negotiations
  • The role of game theory in understanding financial markets
  • The role of game theory in strategic business decisions
  • Game theory and its impact on pricing strategies
  • The use of game theory in understanding competitive strategies
  • Game theory and its role in merger and acquisition decisions
  • The influence of game theory on supply chain management
  • The role of game theory in understanding corporate governance
  • The impact of game theory on marketing strategies
  • Game theory and its role in understanding business negotiations
  • The influence of game theory on organizational behavior
  • The role of game theory in understanding innovation and R&D strategies
  • The role of game theory in understanding voting behavior
  • Game theory and its impact on political negotiations
  • The use of game theory in understanding international relations
  • Game theory and its role in understanding political coalitions
  • The influence of game theory on policy making
  • The role of game theory in understanding electoral competition
  • The impact of game theory on understanding conflict and war
  • Game theory and its role in understanding public policy decisions
  • The influence of game theory on understanding political power dynamics
  • The role of game theory in understanding political cooperation and conflict
  • The role of game theory in algorithm design
  • Game theory and its impact on network design and performance
  • The use of game theory in understanding cybersecurity
  • Game theory and its role in artificial intelligence
  • The influence of game theory on distributed computing
  • The role of game theory in understanding internet economics
  • The impact of game theory on understanding data privacy
  • Game theory and its role in understanding machine learning
  • The influence of game theory on understanding software engineering
  • The role of game theory in understanding social networks
  • The role of game theory in understanding evolutionary biology
  • Game theory and its impact on understanding animal behavior
  • The use of game theory in understanding population dynamics
  • Game theory and its role in understanding cooperation and conflict in nature
  • The influence of game theory on understanding ecological interactions
  • The role of game theory in understanding disease spread and control
  • The impact of game theory on understanding conservation strategies
  • Game theory and its role in understanding genetic algorithms
  • The influence of game theory on understanding ecosystem management
  • The role of game theory in understanding biodiversity
  • The role of game theory in understanding social behavior
  • Game theory and its impact on understanding social norms
  • The use of game theory in understanding social networks
  • Game theory and its role in understanding social conflict and cooperation
  • The influence of game theory on understanding social influence and persuasion
  • The role of game theory in understanding social decision making
  • The impact of game theory on understanding social dilemmas
  • Game theory and its role in understanding social justice and fairness
  • The influence of game theory on understanding social change and stability
  • The role of game theory in understanding social identity and group behavior
  • The role of game theory in understanding mathematical optimization
  • Game theory and its impact on understanding mathematical modeling
  • The use of game theory in understanding mathematical logic
  • Game theory and its role in understanding mathematical decision theory
  • The influence of game theory on understanding mathematical probability
  • The role of game theory in understanding mathematical statistics
  • The impact of game theory on understanding mathematical operations research
  • Game theory and its role in understanding mathematical graph theory
  • The influence of game theory on understanding mathematical combinatorics
  • The role of game theory in understanding mathematical dynamical systems
  • The role of game theory in understanding behavioral economics
  • Game theory and its impact on understanding quantum computing
  • The use of game theory in understanding complex systems
  • Game theory and its role in understanding network science
  • The influence of game theory on understanding artificial intelligence 6. The role of game theory in understanding climate change strategies
  • The impact of game theory on understanding sustainable development
  • Game theory and its role in understanding social media dynamics
  • The influence of game theory on understanding future economic models
  • The role of game theory in understanding future political strategies

This comprehensive list of game theory research paper topics provides a wide range of options for your research. Each category offers unique insights into the different aspects of game theory, from fundamental concepts to future directions. Remember, the best research paper topic is one that not only interests you but also has sufficient resources for you to explore. We hope this list inspires you and aids you in your journey to write a compelling research paper in game theory.

Introduction to Game Theory

Game theory, a fascinating and influential field of study, is a theoretical framework for understanding social situations among competing players. In some respects, game theory is the science of strategy, or at least the optimal decision-making of independent and competing actors in a strategic setting.

The key pioneers of game theory were mathematicians John von Neumann and John Nash, as well as economist Oskar Morgenstern. Game theory was originally developed to understand economic behavior and is still widely used in economics today. However, its applications have expanded to various other fields, including computer science, political science, and biology.

In essence, game theory models the rational behavior of individuals who are aware that their actions affect each other. It involves designing mathematical models to study interactions in structured scenarios (games) and predict the outcome of these interactions. Games in this context are situations where outcomes depend on the actions of multiple players. Each player has a set of possible actions and makes a decision based on their preferences and the expected actions of other players.

Game theory has two main branches: cooperative and non-cooperative game theory. In cooperative games, binding agreements among players are possible, while in non-cooperative games, binding agreements are not. Non-cooperative game theory, which focuses on predicting individual behavior, is more widely used.

Game theory has a wide range of applications. In economics, it is used to model competition and cooperation between firms, to analyze bargaining situations, and to understand and design auctions. In political science, it is used to model strategic voting, while in computer science, it is used in the design of algorithms.

Research papers in game theory allow students to delve deeper into specific areas, contributing to their personal knowledge and the broader academic community. These papers can explore a wide range of game theory research paper topics, from understanding the role of game theory in economic decisions to examining its applications in computer science or political science.

The importance of game theory extends beyond academia. It has real-world implications in various sectors, including economics, politics, and computer science. By understanding the strategic interactions modeled by game theory, we can design better policies, make better business decisions, and develop more efficient algorithms.

How to Choose a Game Theory Topic

Choosing a research topic is a critical step in the research process. The topic you select will guide your study, influence the complexity and relevance of your work, and determine how engaged you are throughout the process. There are numerous intriguing game theory research paper topics that can be explored. Here are some expert tips to assist you in this process:

  • Understanding Your Interests: The first step in choosing a research topic is to understand your interests. What areas of game theory fascinate you the most? Are you interested in how game theory influences economic behavior, or are you more intrigued by its role in computer science or political science? Reflecting on these questions can help you narrow down your options and choose a topic that truly engages you. Remember, research is a time-consuming process, and your interest in the topic will keep you motivated.
  • Evaluating the Scope of the Topic: Once you have identified your areas of interest, the next step is to evaluate the scope of potential game theory research paper topics. A good research topic should be neither too broad nor too narrow. If it’s too broad, you may struggle to cover all aspects of the topic effectively. If it’s too narrow, you may have difficulty finding enough information to support your research. Try to choose a topic that is specific enough to be manageable but broad enough to have sufficient resources.
  • Assessing Available Resources and Data: Before finalizing a topic, it’s important to assess the available resources and data. Are there enough academic sources, such as books, journal articles, and reports, that you can use for your research? Is there accessible data that you can analyze if your research requires it? A preliminary review of literature and data can save you from choosing a topic with limited resources.
  • Considering the Relevance and Applicability of the Topic: Another important factor to consider is the relevance and applicability of the topic. Is the topic relevant to current issues in game theory? Can the findings of your research be applied in real-world settings? Choosing a relevant and applicable topic can increase the impact of your research and make it more interesting for your audience.
  • Seeking Advice: Don’t hesitate to seek advice from your professors, peers, or other experts in the field. They can provide valuable insights, suggest resources, and help you refine your topic. Discussing your ideas with others can also help you see different perspectives and identify potential issues that you may not have considered.
  • Flexibility: Finally, be flexible. Research is a dynamic process, and it’s okay to modify your topic as you delve deeper into your study. You may discover new aspects of the topic that are more interesting or find that some aspects are too challenging to explore due to constraints. Being flexible allows you to adapt your research to these changes and ensure that your study is both feasible and engaging.

Remember, choosing a research topic is not a decision to be taken lightly. It requires careful consideration and planning. However, with these expert tips, you can navigate this process more effectively and choose a game theory research paper topic that not only meets your academic requirements but also fuels your passion for learning.

How to Write a Game Theory Research Paper

Writing a research paper in game theory, like any other academic paper, requires careful planning, thorough research, and meticulous writing. Here are some expert tips to guide you through this process:

  • Understanding the Structure of a Research Paper: A typical research paper includes an introduction, literature review, methodology, results, discussion, and conclusion. The introduction presents your research question and its significance. The literature review provides an overview of existing research related to your topic. The methodology explains how you conducted your research. The results section presents your findings, and the discussion interprets these findings in the context of your research question. Finally, the conclusion summarizes your research and suggests areas for future research.
  • Developing a Strong Thesis Statement: Your thesis statement is the central argument of your research paper. It should be clear, concise, and debatable. A strong thesis statement guides your research and helps your readers understand the purpose of your paper.
  • Conducting Thorough Research: Before you start writing, conduct a thorough review of the literature related to your topic. This will help you understand the current state of research in your area, identify gaps in the literature, and position your research within this context. Use academic databases to find relevant books, journal articles, and other resources. Remember to evaluate the credibility of your sources and take detailed notes to help you when writing.
  • Writing and Revising Drafts: Start writing your research paper by creating an outline based on the structure of a research paper. This will help you organize your thoughts and ensure that you cover all necessary sections. Write a first draft without worrying too much about perfection. Focus on getting your ideas down first. Then, revise your draft to improve clarity, coherence, and argumentation. Make sure each paragraph has a clear topic sentence and supports your thesis statement.
  • Proper Citation and Avoiding Plagiarism: Always cite your sources properly to give credit to the authors whose work you are building upon and to avoid plagiarism. Familiarize yourself with the citation style required by your institution or discipline, such as APA, MLA, Chicago/Turabian, or Harvard. There are many citation tools available online that can help you with this.
  • Seeking Feedback: Don’t hesitate to seek feedback on your drafts from your professors, peers, or writing centers at your institution. They can provide valuable insights and help you improve your paper.
  • Proofreading: Finally, proofread your paper to check for any grammatical errors, typos, or inconsistencies in formatting. A well-written, error-free paper makes a good impression on your readers and enhances the credibility of your research.
  • Incorporating Game Theory Concepts: When writing a game theory research paper, it’s crucial to accurately incorporate game theory concepts. Make sure you understand these concepts thoroughly and can explain them clearly in your paper. Use diagrams and examples where appropriate to illustrate these concepts.
  • Analyzing and Interpreting Game Theory Models: Game theory research often involves analyzing and interpreting game theory models. Be sure to explain your analysis process and interpret the results in a way that is understandable to your readers. Discuss the implications of your findings for the broader field of game theory.
  • Discussing Real-World Applications: Game theory is a practical field with many real-world applications. Discuss how your research relates to these applications. This can make your research more interesting and relevant to your readers.

Remember, writing a research paper is a process that requires time, effort, and patience. Don’t rush through it. Take your time to understand your topic, conduct thorough research, and write carefully. With these expert tips, you can write a compelling game theory research paper that contributes to your academic success and the broader field of game theory.

iResearchNet’s Research Paper Writing Services

In the academic journey, students often encounter challenges that may hinder their ability to produce high-quality research papers. Whether it’s a lack of time, limited understanding of the topic, or difficulties in writing, these challenges can make the process stressful and overwhelming. This is where iResearchNet’s professional writing services come in. We are committed to supporting students in their academic journey by providing top-notch writing services tailored to their unique needs.

  • Expert Degree-Holding Writers: At iResearchNet, we understand the importance of quality in academic writing. That’s why we have a team of expert writers who hold degrees in various fields, including economics. Our writers are not only knowledgeable in their respective fields but also experienced in academic writing. They understand the nuances of writing research papers and are adept at producing well-structured, coherent, and insightful papers.
  • Custom Written Works: We believe that every research paper is unique and should be treated as such. Our writers work closely with you to understand your specific requirements and expectations. They then craft a custom research paper that meets these requirements and reflects your understanding and perspective.
  • In-Depth Research: Good research is the backbone of a compelling research paper. Our writers conduct thorough research using reliable and relevant sources to ensure that your paper is informative and credible. They are skilled at analyzing and synthesizing information, presenting complex ideas clearly, and developing strong arguments.
  • Custom Formatting: Formatting is an essential aspect of academic writing that contributes to the readability and professionalism of your paper. Our writers are familiar with various formatting styles, including APA, MLA, Chicago/Turabian, and Harvard, and can format your paper according to your preferred style.
  • Top Quality: Quality is at the heart of our services. We strive to deliver research papers that are not only well-written and well-researched but also original and plagiarism-free. Our writers adhere to high writing standards, and our quality assurance team reviews each paper to ensure it meets these standards.
  • Customized Solutions: We understand that each student has unique needs and circumstances. Whether you need a research paper on a complex game theory topic, assistance with a specific section of your paper, or editing and proofreading services, we can provide a solution that fits your needs.
  • Flexible Pricing: We believe that professional writing services should be accessible to all students. That’s why we offer flexible pricing options that cater to different budgets. We are transparent about our pricing, and there are no hidden charges.
  • Short Deadlines up to 3 hours: We understand that time is of the essence when it comes to academic assignments. Whether you need a research paper in a week or a few hours, our writers are up to the task. They are skilled at working under pressure and can deliver high-quality papers within short deadlines.
  • Timely Delivery: We respect your deadlines and are committed to delivering your paper on time. Our writers start working on your paper as soon as your order is confirmed, and we keep you updated on the progress of your paper.
  • 24/7 Support: We believe in providing continuous support to our clients. Our customer support team is available 24/7 to answer your questions, address your concerns, and assist you with your order.
  • Absolute Privacy: We respect your privacy and are committed to protecting your personal and financial information. We have robust privacy policies and security measures in place to ensure that your information is safe.
  • Easy Order Tracking: We provide an easy and transparent order tracking system that allows you to monitor the progress of your paper and communicate with your writer.
  • Money Back Guarantee: Your satisfaction is our top priority. If you are not satisfied with our service, we offer a money-back guarantee.

At iResearchNet, we are committed to helping you succeed in your academic journey. We understand the challenges of writing a research paper and are here to support you every step of the way. Whether you need help choosing a topic, conducting research, writing your paper, or editing and proofreading your work, our expert writers are ready to assist you. With our professional writing services, you can focus on learning and leave the stress of writing to us. So why wait? Order a custom game theory research paper from iResearchNet today and experience the difference.

Empower Your Academic Journey: Choose iResearchNet Today!

Embarking on the journey of writing a research paper can be a daunting task, especially when it comes to complex fields like game theory. But remember, you don’t have to face this challenge alone. iResearchNet is here to provide you with the support you need to produce a high-quality, insightful, and impactful research paper.

Our team of expert degree-holding writers is ready to assist you in creating a custom-written research paper that not only meets but exceeds academic standards. Whether you’re struggling with topic selection, research, writing, or formatting, we’ve got you covered. Our comprehensive services are designed to cater to your unique needs and ensure your academic success.

Don’t let the stress of writing a research paper hinder your learning experience. Take advantage of our professional writing services and focus on what truly matters – your learning and growth. With iResearchNet, you can be confident that you’re submitting a top-quality research paper that reflects your understanding and hard work.

So, are you ready to empower your academic journey? Order a custom economics research paper on any of the game theory research paper topics from iResearchNet today. Let us help you navigate your academic journey and secure your success. Remember, your academic achievement is our top priority, and we’re committed to helping you reach your goals. Order now and experience the iResearchNet difference!

ORDER HIGH QUALITY CUSTOM PAPER

research paper about game theory

Information

  • Author Services

Initiatives

You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.

All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .

Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.

Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.

Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.

Original Submission Date Received: .

  • Active Journals
  • Find a Journal
  • Proceedings Series
  • For Authors
  • For Reviewers
  • For Editors
  • For Librarians
  • For Publishers
  • For Societies
  • For Conference Organizers
  • Open Access Policy
  • Institutional Open Access Program
  • Special Issues Guidelines
  • Editorial Process
  • Research and Publication Ethics
  • Article Processing Charges
  • Testimonials
  • Preprints.org
  • SciProfiles
  • Encyclopedia

mathematics-logo

Article Menu

research paper about game theory

  • Subscribe SciFeed
  • Recommended Articles
  • Google Scholar
  • on Google Scholar
  • Table of Contents

Find support for a specific problem in the support section of our website.

Please let us know what you think of our products and services.

Visit our dedicated information section to learn more about MDPI.

JSmol Viewer

An evolutionary game-based regulatory path for algorithmic price discrimination in e-commerce platforms.

research paper about game theory

1. Introduction

2. construction and analysis of evolutionary game models, 2.1. description of the problem, 2.2. underlying assumptions, 2.2.1. strategy space hypothesis, 2.2.2. e-commerce platform-related assumptions, 2.2.3. consumer-related assumptions, 2.2.4. market regulators-related assumptions, 2.2.5. probabilistic assumption, 2.3. model parameter setting, 3. analysis of the two-party evolutionary game, 3.1. the “e-commerce platform-consumer” two-party game, 3.2. “e-commerce platform-market regulators” two-party game, 4. analysis of the three-party evolutionary game, 4.1. game matrix construction, 4.2. strategic stability analysis of three-party game subjects, 4.2.1. e-commerce platform strategy stability analysis, 4.2.2. consumer strategy stability analysis, 4.2.3. analysis of strategic stability in the market regulation sector, 4.3. ess analysis of the subjects of the three-way game, 5. numerical simulation analysis and parametric sensitivity analysis, 5.1. case description, 5.2. analysis of the impact of changes in relevant parameters, 5.2.1. the effect of initial willingness on the outcome of evolutionary games, 5.2.2. the effect of platform penalties on the outcome of evolutionary games, 5.2.3. the effect of consumer advocacy incentives on the outcome of an evolutionary game, 5.2.4. the impact of market regulators’ enforcement acumen on evolutionary game outcomes, 5.2.5. the impact of consumer autonomy likelihood on evolutionary game outcomes, 5.2.6. the effect of punishment of government by superiors on the outcome of the evolutionary game, 5.2.7. the potential benefits of active supervision by market regulators on the outcome of evolutionary games, 6. conclusions and discussion, 6.1. conclusions, 6.2. suggestion, 6.2.1. consumers, 6.2.2. market regulator, 6.3. limitations, author contributions, data availability statement, acknowledgments, conflicts of interest.

  • Islek, I.; Oguducu, S.G. A hierarchical recommendation system for E-commerce using online user reviews. Electron. Commer. Res. Appl. 2022 , 52 , 101131. [ Google Scholar ] [ CrossRef ]
  • Gonçalves, M.J.A.; Pereira, R.H.; Coelho, M.A.G.M. User Reputation on E-Commerce: Blockchain-Based Approaches. J. Cybersecur. Priv. 2022 , 2 , 907–923. [ Google Scholar ] [ CrossRef ]
  • Abrardi, L.; Cambini, C.; Rondi, L. Artificial intelligence, firms and consumer behavior: A survey. J. Econ. Surv. 2022 , 36 , 969–991. [ Google Scholar ] [ CrossRef ]
  • Keller, A.; Vogelsang, M.; Totzek, D. How displaying price discounts can mitigate negative customer reactions to dynamic pricing. J. Bus. Res. 2022 , 148 , 277–291. [ Google Scholar ] [ CrossRef ]
  • Wang, J.; Shu, T.; Zhao, W.; Zhou, J. Research on Chinese Consumers’ Attitudes Analysis of Big-Data Driven Price Discrimination Based on Machine Learning. Front. Psychol. 2022 , 12 , 803212. [ Google Scholar ] [ CrossRef ]
  • Zhu, Z. Legal Regulation of Algorithmic Discrimination. Adv. Soc. Behav. Res. 2021 , 1 , 65–72. [ Google Scholar ] [ CrossRef ]
  • Wang, F. China’s Regulatory Framework for Dynamic and Personalized Pricing in the Digital Economy. Int. J. Soc. Sci. Stud. 2022 , 10 , 1–10. [ Google Scholar ] [ CrossRef ]
  • Zhao, Q. Exploration on the protection of consumer rights and interests in big data price discrimination. Int. J. Front. Sociol. 2024 , 6 , 060403. [ Google Scholar ]
  • Li, Y. The Anti-nonopoly Law Regulation of Big Data Killing in the Background of Digital Economy. J. Econ. Law 2024 , 1 , 214–222. [ Google Scholar ] [ CrossRef ]
  • Hu, Z. Liability Rules for Algorithmic Discrimination. Sci. Law J. 2023 , 2 , 020902. [ Google Scholar ]
  • Wang, Y. Price Discrimination in the Era of Big Data ; Atlantis Press: Amsterdam, The Netherlands, 2022; 664p. [ Google Scholar ]
  • Yang, X.; Dai, X.; Bin, H. The Dynamics of Rewards and Penalties: Governmental Impact on Green Packaging Adoption in Logistics. Sustainability 2024 , 16 , 4835. [ Google Scholar ] [ CrossRef ]
  • Hu, Z.; Wang, Y.; Zhang, H.; Liao, W.; Tao, T. An evolutionary game study on the collaborative governance of environmental pollution: From the perspective of regulatory capture. Front. Public Health 2024 , 11 , 1320072. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Luo, E.; Xiang, S.; Yang, Y.; Sethi, N. A stochastic and time-delay evolutionary game of food safety regulation under central government A stochastic and time-delay evolutionary game of food safety regulation under central government punishment mechanism. Heliyon 2024 , 10 , e30126. [ Google Scholar ]
  • Liu, Z.; Zhao, L.; Wu, C. Supervision Timing Simulation Analysis of Community E-commerce Platform Supply Chain Based on Tripartite Game Model. Teh. Vjesn. 2022 , 29 , 2008–2017. [ Google Scholar ]
  • Liu, X.; Huang, Z.; Wang, Q.; Wan, B. An Evolutionary Game Theory-Based Method to Mitigate Block Withholding Attack in Blockchain System. Electronics 2023 , 12 , 2808. [ Google Scholar ] [ CrossRef ]
  • Pan, K.; Wang, L.; Zhang, L. A Study on Enhancing the Information Security of Urban Traffic Control Systems Using Evolutionary Game Theory. Electronics 2023 , 12 , 4856. [ Google Scholar ] [ CrossRef ]
  • Friedman, D. On economic applications of evolutionary game theory. J. Evol. Econ. 1998 , 8 , 15–43. [ Google Scholar ] [ CrossRef ]
  • Luo, J.; Huang, M.; Bai, Y. Promoting green development of agriculture based on low-carbon policies and green preferences: An evolutionary game analysis. Environ. Dev. Sustain. 2023 , 26 , 6443–6470. [ Google Scholar ] [ CrossRef ]
  • Bai, S.; Liu, Z.; Lv, Y. Evolutionary Game Analysis of Consumer Complaint Handling in E-Commerce. Discret. Dyn. Nat. Soc. 2022 , 2022 , 3792080. [ Google Scholar ] [ CrossRef ]
  • Zha, Y.; Wang, Y.; Li, Q.; Yao, W. Credit offering strategy and dynamic pricing in the presence of consumer strategic behavior. Eur. J. Oper. Res. 2022 , 303 , 753–766. [ Google Scholar ] [ CrossRef ]
  • Yu, D.; Li, X. Dynamic evolutionary game research of big data-based price discrimination between consumers and merchants. Price Theory Pract. 2019 , 11 , 129–132. [ Google Scholar ]
  • Xing, G.; Lu, F.; Luo, D. An Evolutionary Simulation Analysis of E-Commerce Big Data-Based Price Discrimination under Government Supervision. J. Hunan Univ. Technol. 2021 , 35 , 65–72. [ Google Scholar ]
  • Qiu, Z.; Yin, Y.; Yuan, Y.; Chen, Y. Research on Credit Regulation Mechanism of E-commerce Platform Based on Evolutionary Game Theory. J. Syst. Sci. Syst. Eng. 2024 , 33 , 330–359. [ Google Scholar ] [ CrossRef ]
  • Statistical Mechanics; Data on Statistical Mechanics Reported by Researchers at Nanjing Tech University (Tripartite Evolutionary Game Analysis for “deceive Acquaintances” Behavior of E-commerce Platforms in Cooperative Supervision). J. Phys. Res. 2020 .
  • Li, J.; Xu, X.; Yang, Y. Research on the Regulation of Algorithmic Price Discrimination Behaviour of E-Commerce Platform Based on Tripartite Evolutionary Game. Sustainability 2023 , 15 , 8294. [ Google Scholar ] [ CrossRef ]
  • Chen, Y.-F.; Pang, T.-T.; Kuslina, B.H. The Effect of Price Discrimination on Fairness Perception and Online Hotel Reservation Intention. J. Theor. Appl. Electron. Commer. Res. 2023 , 18 , 1320–1337. [ Google Scholar ] [ CrossRef ]
  • Wu, Z.; Yang, Y.; Zhao, J.; Wu, Y. The Impact of Algorithmic Price Discrimination on Consumers’ Perceived Betrayal. Front. Psychol. 2022 , 13 , 825420. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Hu, L.; Zhang, Y.; Chung, S.-H.; Wang, L. Two-tier price membership mechanism design based on user profiles. Electron. Commer. Res. Appl. 2022 , 52 , 101130. [ Google Scholar ] [ CrossRef ]
  • Peiseler, F.; Rasch, A.; Shekhar, S. Imperfect information, algorithmic price discrimination, and collusion. Scand. J. Econ. 2022 , 124 , 516–549. [ Google Scholar ] [ CrossRef ]
  • Liu, W.; Long, S.; Xie, D.; Liang, Y.; Wang, J. How to govern the big data discriminatory pricing behavior in the platform service supply chain? An examination with a three-party evolutionary game model. Int. J. Prod. Econ. 2020 , 231 , 107910. [ Google Scholar ] [ CrossRef ]
  • Aparicio, D.; Misra, K. Artificial Intelligence and Pricing ; Emerald Publishing Limited: Bradford, UK, 2023; Volume 20, pp. 103–124. [ Google Scholar ]
  • van der Rest, J.-P.; Sears, A.M.; Kuokkanen, H.; Heidary, K. Algorithmic pricing in hospitality and tourism: Call for research on ethics, consumer backlash and CSR. J. Hosp. Tour. Insights 2022 , 5 , 771–781. [ Google Scholar ] [ CrossRef ]
  • Deng, C.; Zhu, H. On the Legal Status of OTA Network Platform: Taking the “Ctrip Price Discrimination” Case as an Example. J. Hunan Univ. Technol. (Soc. Sci. Ed.) 2023 , 28 , 60–71. [ Google Scholar ]
  • Zhou, K.; Wang, Q.; Tang, J. Evolutionary game and simulation analysis of enterprise’s green technology innovation under green credit policy: Evidence from China. Technol. Anal. Strateg. Manag. 2024 , 36 , 1551–1570. [ Google Scholar ] [ CrossRef ]

Click here to enlarge figure

ParametersSignificance
xProbability that the e-commerce platform will adopt normal pricing
yProbability of a consumer advocacy report
zProbability of active government regulation
C The cost of regulating the platform’s operations
C Costs when platforms adopt algorithmic price discrimination
C Costs when governments adopt active supervision
C Costs to consumers when reporting
P Pricing when platforms adopt regulated operations
P Pricing when platforms adopt algorithmic price discrimination
L Loss of platform reputation when platform algorithmic price discrimination occurs in consumer advocacy reports
L Loss of social credibility when platform algorithmic price discrimination occurs with negative government regulation
R Potential benefits when actively regulated by the government (increased social credibility, rewards from higher authorities)
F Government fines for platforms that engage in algorithmic price discrimination
F Penalties imposed by superiors when platforms algorithmically price discriminate, and governments passively regulate them
V Consumers’ estimated value of platform goods and services
R Psychological gains when consumers compare prices among themselves and find no algorithmic price discrimination by platforms
R Government incentives when platform algorithmic price discrimination occurs in consumer advocacy reports
R Potential benefits of increased consumer trust for e-commerce platforms
When the government actively regulates, the acuity of the platforms’ algorithmic price discrimination is found to increase to
Consumers who find out that they have been subjected to price discrimination have a likelihood of defending their rights to recover the difference in price
Consumers
Reporting
Acceptance
E-commerce PlatformNormal Pricing
Discriminatory Pricing

Balance Point






Market Regulator
Active Supervision
Passive Supervision
E-commerce PlatformNormal Pricing
, ,
Discriminatory Pricing
,
,
Balance Point



ConsumersMarket Regulator
Active Supervision
Passive Supervision
E-commerce platformNormal Pricing
Reporting




Acceptance




Discriminatory Pricing
Reporting




Acceptance




Balance PointHidden Meaning
Discriminatory pricing by e-commerce platforms; consumers tolerate and accept; passively regulated by market regulators
Discriminatory pricing by e-commerce platforms; consumers tolerate and accept; actively regulated by market regulators
Discriminatory pricing by e-commerce platforms; consumers report; active supervision by market regulators
Discriminatory pricing by e-commerce platforms; consumers report; passive supervision by market regulators
E-commerce platforms regulate pricing; consumers tolerate and accept; market regulators passively regulate
E-commerce platforms regulate pricing; consumers tolerate and accept; market regulators passively regulate
E-commerce platforms standardize pricing; consumers report; market regulators actively regulate
E-commerce platforms standardize pricing; consumers report; market regulators actively regulate

Balance Point
NN
NN
NN
NN
NN
NNN
NNN
N N
Note: N means uncertainty.
ParametersValueSource (Reference)
C 5Principle of Inequality
C 10
C 8CEIC Data
Salarylnsights
C 16Scheme for the Payment of Litigation Costs
P 20Deng Chunmei and Zhu Heng [ ]
P 40
F 10
R 5
L 8Nanjing Tech [ ]
Jianjun Li et al. [ ]
L 5
R 12
R 3
R 3
F 5Principle of Inequality
α0.2Zhou Kui et al. [ ]
β0.2
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

Guo, Y.; Lin, J.; Zhuang, W. An Evolutionary Game-Based Regulatory Path for Algorithmic Price Discrimination in E-Commerce Platforms. Mathematics 2024 , 12 , 2774. https://doi.org/10.3390/math12172774

Guo Y, Lin J, Zhuang W. An Evolutionary Game-Based Regulatory Path for Algorithmic Price Discrimination in E-Commerce Platforms. Mathematics . 2024; 12(17):2774. https://doi.org/10.3390/math12172774

Guo, Yan, Jiajun Lin, and Weiqing Zhuang. 2024. "An Evolutionary Game-Based Regulatory Path for Algorithmic Price Discrimination in E-Commerce Platforms" Mathematics 12, no. 17: 2774. https://doi.org/10.3390/math12172774

Article Metrics

Article access statistics, further information, mdpi initiatives, follow mdpi.

MDPI

Subscribe to receive issue release notifications and newsletters from MDPI journals

IMAGES

  1. (PDF) Game theory in marketing science uses and limitations

    research paper about game theory

  2. Game Theory Research Paper Example

    research paper about game theory

  3. Game Theory

    research paper about game theory

  4. Game Theory: Assumptions, Application and Limitations

    research paper about game theory

  5. (PDF) A Review of Game Theory Application Research in Safety Management

    research paper about game theory

  6. (PDF) Game Theory, Introduction to

    research paper about game theory

VIDEO

  1. 05 01 Introduction to Game Theory

  2. Game Theory

  3. Game-theoretic probability and its applications

  4. Operation Research

  5. Lecture 46: Evolutionary Game Theory- A Deviation from Regular Game Theory

  6. Practical Game Theory

COMMENTS

  1. An Overview on Game Theory and Its Application

    This research paper looks into Game Theory while immensely addressing the historical background of this theory. The document also gives an overview and definition of relevant terminologies related ...

  2. Home

    International Journal of Game Theory

  3. (PDF) Game Theory

    Game Theory (GT) is the formal study of conflict and cooperation among sev-. eral agents, denoted as players, representing individuals, animals, computers, groups, firms, etc. The concepts of ...

  4. PDF Game theory and applications

    interaction and collective-level dynamics are uni˙ed by means of game theory. These papers do also show the generality of applications. The results are abstract and not restricted to a certain context or interpretation. 1.1 The early days of game theory We begin this chapter with one of the most important results of game

  5. Articles

    Jan-Peter Siedlarek. Original Paper 10 May 2023 Pages: 1037 - 1076. 1. 2. …. 37. Next. International Journal of Game Theory is a renowned scholarly publication dedicated to game theory and its applications. Publishes research with significant ...

  6. Applications of game theory in deep learning: a survey

    This paper provides a comprehensive overview of the applications of game theory in deep learning. Today, deep learning is a fast-evolving area for research in the domain of artificial intelligence. Alternatively, game theory has been showing its multi-dimensional applications in the last few decades. The application of game theory to deep learning includes another dimension in research. Game ...

  7. Game Theory and Applications in Economics: Journal of Applied Mathematics

    Game theory, as a branch of the applied mathematics, contributes insight into economics, sociology, and many other disciplines. Game theory has outstanding effects on the theory of economics, and the literature on applying game theoretical and related approaches to economics is growing rapidly, and yet many theoretical and empirical challenges remain in this field.

  8. Scientific & Academic Publishing: Aims and Scope

    The Journal of Game Theory publishes original research, survey papers and short notes in game theory and its applications providing rigorous analysis and significant contributions from a methodological, conceptual or mathematical point of view. It welcomes game-oriented papers from all fields of study, including: biology; computer science ...

  9. Game Theory and Applications

    Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications. ... Game theory, which is the study ...

  10. 100781 PDFs

    Explore the latest full-text research PDFs, articles, conference papers, preprints and more on GAME THEORY. Find methods information, sources, references or conduct a literature review on GAME THEORY

  11. An Overview on Game Theory and Its Application

    This research paper looks into Game Theory while immensely addressing the historical background of this theory. The document also gives an overview and definition of relevant terminologies related to this theory like a game, Nash equilibrium, and dominance which form the basis of the theory concept. It also focuses on mixed strategies ...

  12. PDF Introduction to Game Theory

    Title. Introduction to Game Theory. Author. Jonathan Cave. Subject. This paper, originally written as a pedagogical note for a RAND Graduate School course, introduces the concepts and tools of game theory in the context of microeconomics. Created Date. 11/21/2008 1:45:11 PM.

  13. THINKING POKER THROUGH GAME THEORY

    the title of this project suggest, a grasp on game theory. Game theory has many applica-tions in a wide array of elds such as economics, politics, psychology, science, computer science, and biology to name a few. Although, we could spend an entire project studying game theory alone, we will concentrate on decision making and solving 2-person games.

  14. Special Issue on Game Theory

    The 13 papers in this special issue focus on game theory. The aim of this issue is to bring together the state-of-the-art research contributions that address the fundamentals and sound theoretical models of game theory, and the major opportunities and challenges of applying game theory to solving real problems in industry, biology, medicine ...

  15. PDF Computer Science and Game Theory

    computer scientists, the paper took stock of the interaction among game theory, operations research and computer science at the time.2 The paper points to the following areas: 1. Graphs in games. 2. The complexity of solving a game. 3. Multi-person operations research. 4. The complexity of playing a game. 5. Modelling bounded rationality.

  16. PDF The Work of John Nash in Game Theory

    H ARSANYI. In the short period of 1950 - 53, John Nash published four brilliant papers ([35], [37], [38], [39]), in which he made at least three fundamentally important contributions to game theory: (1) He introduced the distinction between cooperative and non-cooperative games.

  17. PDF Game Theory

    Research in these applications of game theory is the topic of recent conference and journal papers (see, for example, Binmore and Vulkan, "Applying game theory to auto-mated negotiation," Netnomics Vol. 1, 1999, pages 1-9) but is still in a nascent stage. The

  18. Recent Developments of Game Theory and Reinforcement Learning

    The reviewed papers mainly focus on broad themes and address three important research questions: the impact of game theory on multi-agent reinforcement learning (MARL), the significant ...

  19. Game theory applications in production research in the sharing and

    Finally, the last paper belongs to the section of circular economy is by Qin et al. (Citation 2019) who study carbon emissions in a capital-constrained manufacturer in supply chains and the paper's title is 'The Value of Advance Payment Financing to Carbon Emission Reduction and Production in a Supply Chain with Game Theory Analysis'. The ...

  20. Recent Developments of Game Theory and Reinforcement Learning

    In the ever-changing world of decision-making, when game theory and reinforcement learning(RL) come together, they create a fascinating combination that shows a new way to solve complex problems in many fields. The combination of game theory and RL is a powerful convergence that opens up a hopeful new frontier for dealing with complex decision-making problems in many different fields. Research ...

  21. Scientific & Academic Publishing: Articles

    The Journal of Game Theory publishes original research, survey papers and short notes in game theory and its applications providing rigorous analysis and significant contributions from a methodological, conceptual or mathematical point of view. It welcomes game-oriented papers from all fields of study, including: biology; computer science ...

  22. A Brief Introduction to the Basics of Game Theory

    Abstract. I provide a (very) brief introduction to game theory. I have developed these notes to provide quick access to some of the basics of game theory; mainly as an aid for students in courses in which I assumed familiarity with game theory but did not require it as a prerequisite. Keywords: Game Theory. undefined.

  23. How to address enterprise collusion in ...

    Addressing this issue requires a comprehensive strategy involving various stakeholders. This paper focuses on collusive behavior in carbon emissions data falsification. An evolutionary game model is established to illustrate the interactions between local governments, carbon-emitting enterprises, and third-party carbon verification agencies.

  24. Game Theory Research Paper Topics

    This comprehensive guide to game theory research paper topics is designed to assist students and researchers in the field of economics. Selecting a compelling and relevant research topic is a crucial step in the academic journey, and this guide aims to facilitate this process. We provide an extensive list of topics, divided into ten categories ...

  25. An Evolutionary Game-Based Regulatory Path for Algorithmic Price ...

    To effectively regulate this behavior, the paper utilizes evolutionary game theory (EGT) to analyze the strategies employed by e-commerce platforms, consumers, and market regulators to achieve stability. ... Yu, D.; Li, X. Dynamic evolutionary game research of big data-based price discrimination between consumers and merchants. Price Theory ...

  26. Dynamic environment UAV deployment algorithm based on potential game theory

    This paper centers on the data broadcast scenario of ground sensor nodes, addressing the problem of using multiple UAVs to cover large regional areas. It constructs a distributed game theory model and proposes an exploratory spatial adaptive play algorithm to solve Nash equilibrium in the game process [16, 17]. The main content and ...