1. Introduction
The
Distributed Constraint Optimization Problems (DCOPs) [
1,
2,
3,
4] are an efficient framework to solve the cooperative problems of multi-agent systems. In DCOPs, agents communicate directly or indirectly to coordinate their value assignments of variables to optimize the aggregate constraint utilities, where the variables are discrete and the constraint utilities are presented in tabular form. DCOPs have been successfully applied to model the problems in the real world like distributed meeting scheduling [
5,
6], sensor networks [
7], multi-robot coordination [
8,
9], smart grid [
10,
11], and smart home [
12,
13]. Some improved methods such as [
14,
15,
16,
17] and some improvements of key metrics like privacy [
18,
19,
20] have been proposed. In addition, there are some extensions of DCOPs like Asymmetric DCOPs [
21,
22,
23,
24] and Dynamic DCOPs [
25,
26,
27].
Over the years, researchers have proposed several different types of algorithms to solve DCOPs, including exact search-based algorithms [
28,
29], exact inference-based algorithms [
2,
30], approximate search-based algorithms [
31], approximate inference-based algorithms [
7], and approximate sampling-based algorithms [
32,
33]. Although the DCOP algorithms have excellent performance for discrete problems, it is not satisfactory for solving continuous problems, such as target tracking sensor orientation. Therefore,
Continuous DCOPs (C-DCOPs) [
34] were proposed to model problems with continuous variables, and the constraint utilities are functional forms. Correspondingly, researchers proposed new C-DCOP algorithms to deal with the modification of the C-DCOP formulation, including
Continuous Max-Sum (CMS) [
34],
Hybrid CMS (HCMS) [
35],
Bayesian Distributed Pseudo-tree Optimization Procedure (B-DPOP) [
36],
Particle Swarm Based Continuous DCOP (PCD) [
37],
Exact Continuous DPOP (EC-DPOP) [
38] and extensions (
Approximate Continuous DPOP (AC-DPOP),
Clustered AC-DPOP (CAC-DPOP), and
Continuous DSA (C-DSA)),
Continuous Cooperative Constraint Approximation (C-CoCoA) [
39], and
Particle Swarm with Local Decision Based Continuous DCOP (PCD-LD) [
40].
The CMS algorithm is a continuous extension of discrete Max-Sum (MS) algorithm, which approximates the utility functions as piece-wise linear functions to solve C-DCOPs. Similarly, HCMS improves the solution quality according to the non-linear optimization method based on the MS algorithm. Later, B-DPOP combines the Bayesian Optimization framework with DPOP to solve C-DCOPs. After that, Hoang et al. propose three extensions of the inference-based DPOP algorithm and one extension of search-based Distributed Stochastic Algorithm (DSA): EC-DPOP, AC-DPOP, CAC-DPOP, and C-DSA, where the first algorithm provides an exact approach to solve general C-DCOPs, and the second and third provide approximate methods. In addition, C-DSA extends DSA to the continuous domain. C-CoCoA is a semi-greedy non-iterative algorithm that can get a high-quality solution with less execution time and overhead.
Unfortunately, none of the above C-DCOP algorithms is an anytime algorithm. An anytime algorithm should have two characteristics:
The algorithm can be stopped at any time to output a solution, which means it is iterative;
The algorithm guarantees the monotonicity of the solution quality over time. PCD introduces the
Particle Swarm Optimization (PSO) algorithm to solve C-DCOPs, and PCD-LD combines local decisions with PCD to improve the solution quality. They are two iterative algorithms and collect all utilities by constructing a
Breadth First Search (BFS) pseudo-tree to guarantee the monotonicity of the solution quality. Although PCD and PCD-LD are anytime algorithms, the utility passing in the pseudo-tree causes privacy violations [
41].
In the wake of the above analysis, we extend
Maximum Gain Message (MGM) [
31] algorithm so that it’s able to solve C-DCOPs. Three main reasons prompt us to customize MGM for C-DCOPs as follows:
We consider that anytime property is an important problem concerned for the C-DCOP algorithm. Since communication may be halted arbitrarily, an algorithm without the anytime property risk being terminated at an unsatisfactory assignment combination. Therefore, an anytime algorithm guarantees lower bounds on performance in anytime environments when given acceptable starting conditions.
The MGM algorithm is an iterative, search-based algorithm that performs a distributed local search, and it can guarantee the monotonicity of the solution quality through local interactions. In other words, the MGM algorithm is an anytime algorithm.
Compared with the anytime algorithms using the BFS pseudo-tree, the MGM algorithm solves problems without any restriction on the graph structure. Specifically, MGM uses a basic constraint graph and local interactions to maintain the privacy of agents.
Therefore, we provide three MGM-based C-DCOP algorithms, namely Continuous MGM (C-MGM), Parallel C-MGM (C-PMGM), and Parallel Differential Search C-MGM (C-PDSM). Firstly, C-MGM is an approximation C-DCOP algorithm that is similar to MGM. In C-MGM, each agent compares the gain of its local utility with neighbors to strictly increase the aggregate utility. Then, C-PMGM extends C-MGM and generates a competing solution for each solution to solve C-DCOPs in parallel. Finally, C-PDSM is an improvement based on C-PMGM that uses differential search to update competing solutions. We empirically evaluate the performance of proposed algorithms on benchmark problems, and the results show that our algorithms outperform the state-of-the-art C-DCOP algorithms.
The remainder of this paper is organized as follows. The next section introduces the necessary background, including the DCOP and C-DCOP frameworks, and the standard MGM algorithm. The details of our algorithms are described in
Section 3. Then, we present theoretical analysis in
Section 4. Afterward,
Section 5 reports the empirical evaluation of our algorithms against the state-of-the-art algorithms. Finally, we provide conclusions and future work in
Section 6.
2. Background
In this section, we provide the definition of the DCOP framework. Then, we introduce the C-DCOP framework and provide an example of a C-DCOP. Finally, we describe the MGM algorithm, which is the basis of our proposed algorithms.
2.1. Distributed Constraint Optimization Problems
A Distributed Constraint Optimization Problem (DCOP) is a tuple , where
is a set of agents, an agent can control one or more variables.
is a set of discrete variables, each variable is controlled by one of the agents.
is a set of discrete domains and each variable takes value from the set .
is a set of utility functions and each utility function is defined over a set of variables: , where the is the scope of .
is a mapping function that associates each variable to one agent. In this paper, we assume one agent controls only one variable.
In this paper, we assume that one agent only controls one variable, thus and the terms "agent" and "variable" could be used interchangeably. In addition, we assume that all utility functions are binary functions between two variables, thus and the .
A solution of a DCOP is an assignment combination
to all variables that are consistent with their respective domains. The goal of the solution is to minimize
1 the aggregate utility as shown in Equation (
1), where the aggregate utility is the sum of utilities across all the applicable utility functions.
2.2. Continuous Distributed Constraint Optimization Problems
The Continuous DCOP (C-DCOP) framework extends the DCOP framework by modeling the variables as continuous variables. It is defined by a tuple , where , , and are the same as DCOP. The set of variables and the set of domains are defined as follows:
is the set of continuous variables.
is the set of continuous domains and each continuous variable takes any value from the domain , where and represent the lower and upper bounds of the domain, respectively.
The goal of a C-DCOP is the same as Equation (
1), where the variables
x and domain
are continuous.
Figure 1 provides an example of a C-DCOP, where
Figure 1(a) is a constraint graph with three variables controlled by the agents and three constraints.
Figure 1(b) defines the utility functions corresponding to the constraints and gives the domain of variables.
2.3. Maximum Gain Message
The MGM algorithm is an anytime, local search algorithm for solving DCOPs. Its main idea is to guarantee the agent with maximum gain changes its variable through competition between neighboring agents. In this decision method, only one agent can change its value in each local, which can completely avoid the uncertainty of the result caused by the simultaneous assignment of neighboring agents. Therefore, the solution quality improves steadily with iterations.
The framework of MGM is presented in Algorithm 1. Firstly, each agent randomly selects a value and sends it to neighbors (Algorithm 1: Lines 2-3). After receiving these values from neighbors, the agent assumes that the neighbors’ values remain unchanged and chooses a new value that maximizes the local gain. Then, the agent sends the local gain to neighbors and simultaneously receives gains from neighbors. If the agent with the maximum gain in the current local, it’s assigned the new value, otherwise it remains the chosen random value. Finally, the agent sends the updated value to neighbors and executes the loop (Algorithm 1: Lines 5-12) until the condition is met.
Algorithm 1: Maximum Gain Message Algorithm (MGM) |
|
3. Our Algorithms
In this section, we introduce the three proposed algorithms in detail. Let us define a
solution set containing
K solutions, where
represents
an assignment combination (i.e. a solution) of
n variables controlled by agents. Therefore, an agent
holds the value assignments of
K solutions on the
dimension, denoted as
. In addition, we use
to represent the variable of the
solution held by agent
, and its local utility is
, where
represents two variables on the
solution and
represents the neighbors of agent
. For example, we construct a solution set with two solutions for
Figure 1 as
, where
and
. For agent
, its value assignments are
and the local utility of
is
. It is worth noting that since we consider minimizing the aggregate utility, we use
absolute gain to represent the local utility reduced, which is a negative value.
3.1. Continuous Maximum Gain Message
Algorithm 2 shows the process of the
Continuous Maximum Gain Message (C-MGM) algorithm, including three phases:
Initialization,
Calculation, and
Comparison.
Algorithm 2: Continuous MGM (C-MGM) |
|
Initialization: Firstly, C-MGM constructs a solution set for K solutions. Then, each agent assigns a random value from its domain to each solution on the current dimension and generates a variable to record the gain value of agent . Finally, the agent sends ValueMessage() to all neighbors (Algorithm 2: Lines 1-7), where is a random value in .
Calculation: After receiving ValueMessage() from neighbors, the agent calculates the local utility for each solution according to its value assignments , respectively (Algorithm 2: Lines 12-14). We define to represent the basic local utility of agent , which is the local utility assuming that the values of agent and its neighboring agents are not changed, calculated by . Then, the is assigned a difference value between the minimum and basic local utility (Algorithm 2: Lines 15-16). Finally, the agent sends GainMessage() to neighbors.
Comparison: Similarly, the agent enters the comparison phase after receiving GainMessage() from each neighbor . If the current agent with the minimum gain value among all gain values of neighbors and the gain is less than 0, the agent assigns the value that minimizes local utility to (Algorithm 2: Lines 18-20). In general, after the size of the domain is determined, the variable of the standard MGM algorithm can take a finite number of values from its domain, but there is an infinite number of values to be selected in the variable of C-MGM. We consider updating the value assignments for each agent to search its domain and avoid falling in the local optimum by adding randomness. Therefore, the agent updates each value to a random value from its domain (Algorithm 2: Line 22).
3.2. Parallel C-MGM
We consider solving
K solutions in parallel based on C-MGM. Therefore, we propose the
Parallel C-MGM (C-PMGM) algorithm as shown in Algorithm 3, which also consists of three phases:
Initialization,
Calculation, and
Comparison.
Algorithm 3: Parallel C-MGM (C-PMGM) |
|
Initialization: Except for the similar initialization as C-MGM, C-PMGM additionally generates a competing solution set , which is the same as the solution set . In addition, each agent uses a gain set to record the gain values of K solutions (Algorithm 3: Lines 7-9). It is worth noting that the ValueMessage() sent by agent in C-PMGM to neighbors is the value assignments instead of a random value in (Algorithm 3: Line 10).
Calculation: Since the computational overhead of solving K solutions according to the way of C-MGM is expensive, we consider using the values from the corresponding competing solution set to calculate the gains for K solutions (Algorithm 3: Lines 11-16), which means that the indexes of the two solution sets ( and ) are consistent. Later, the agent sends GainMessage() to neighbors.
Comparison: Similar to C-MGM, the agent in C-PMGM compares K gains with all neighbors. For each gain in the gain set of agent , if the gain value is the smallest in the current local and less than 0, the value is assigned to its competing value (Algorithm 3: Lines 18-21). Since an agent only calculates one local utility for a solution instead of K local utilities (such as Algorithm 2: Lines 13-14), the idea is to search as many competing values as possible. Therefore, the agent resets each competing value to a value from its domain, whether or not the value is changed.
3.3. Parallel Differential Search C-MGM
Based on C-PMGM, we construct a heuristic method to guide the exploitation of the competing solution set. Specifically, we extend the update of value assignments in C-PMGM by combining a differential search, named
Parallel Differential Search C-MGM (C-PDSM). The details of C-PDSM can be found in Algorithm 4.
Algorithm 4: Parallel Differential Search C-MGM (C-PDSM) |
|
The difference between C-PDSM and C-PMGM is the update of value assignments (Algorithm 4: Line 22). For a gain set in one agent, the agent filters each absolute gain and adds it to an
absolute gain set (Algorithm 4: Lines 26-29). Since the differential search method we used requires at least two absolute gains, we provide two ways to update value assignments (Algorithm 4: Lines 30-34):
If the number of absolute gains is less than two, the agent update the value assignments of the competing solution set through random search.
Otherwise, the agent updates value assignments by differential search, as shown in Equation (
2). Our motivation is that the current value learns the exploitation ability from other values with absolute gain, which can improve the convergence speed of the algorithm.
where
is a scaling factor
, which controls the amplification of the differential variation [
42]. The "
" and "
" represent the indexes of competing solution set with the minimum and maximum absolute gain values, respectively.
3.4. An Example of Algorithms
We take
Figure 1 as an example to realize the three phases of C-MGN, C-PMGM, and C-PDSM.
In the Initialization phase, we assume that the size of the solution set is and the solution sets of the three algorithms are all . Therefore, the value assignments held by agent are , the value assignments of agent are , and the value assignments of agent are .
For C-MGM, the values randomly selected by each agent are , , and .
For C-PMGM and C-PDSM, we assume the competing solution sets both are and the scaling factor in C-PDSM is 1.4.
In the Calculation phase, the agent calculates the local utility and gain after receiving the random value or value assignments from neighbors . Let’s take the agent as an example for calculation.
-
For C-MGM:
;
-
For C-PMGM and C-PDSM:
;
;
In the Comparison phase, each agent compares the gains to action. Simultaneously, there are different way to update value assignments in C-MGM, C-PMGM, and C-PDSM. We assume that the agent with maximum gain among its neighbors, and its actions and updates are as follows:
-
For C-MGM:
;
(assumed random values);
-
For C-PMGM:
, ;
(assumed random values);
-
For C-PDSM:
, ;
Since both gain values (-189 and -43) are less than 0, .
;
Finally, C-MGM outputs a solution as . The two solutions output by C-PMGM and C-PDSM are and , the difference between C-PMGM and C-PDSM is that the value assignments of agent in the competing solution set are and , respectively.
4. Theoretical Analysis
In this section, we prove that C-MGM, C-PMGM, and C-PDSM are three anytime algorithms, and we provide some theoretical properties of our algorithms in terms of communication, computation, and memory. We assume that the graph structure is a binary constraint graph , the number of nodes is , and the number of edges in the constraint graph is . In addition, we define one iteration as the complete process of the algorithm completing Calculation and Comparison phase.
Theorem 1. C-MGM, C-PMGM, and C-PSDM are three anytime algorithms.
Proof. On the one hand, the three algorithms are iterative, which means they can be terminated at any time and output solutions. On the other hand, we assume that the aggregate utility (U) of C-MGM at t iteration is , which represents the sum of utility on all constraints. When an agent calculates local utility, the same constraint is calculated twice by two neighboring agents, the aggregate utility can be modified as . Therefore, the aggregate utility of the iteration in C-MGM can represent , where represents a set of variables that change their values. In each local, since only one variable with local utility reduced changes its value, we can get that . Therefore, , which means the aggregate utility strictly decreases with increasing iterations. Similarly, C-PMGM and C-PDSM solve K solutions in parallel, the aggregate utilities’ change of one solution in them is the same as that of C-MGM. Hence, C-MGM, C-PMGM, and C-PSDM are three anytime algorithms. □
Theorem 2. The total number of messages of C-MGM, C-PMGM, and C-PDSM with t iterations all are .
Proof. For each constraint in one iteration, the agents of the three algorithms need to send value assignments to neighbors in the Calculation phase and vice versa, there are two messages per edge. Therefore, the number of messages is in this phase, where is the number of constraints (edges). In the Comparison phase, the agents of the three algorithms share gain or gain sets with neighbors to compare their values, thus the number of messages is also . Finally, the total number of messages of C-MGM, C-PMGM, and C-PDSM with t iterations all are . □
Theorem 3. In one iteration, the message size of C-MGM, C-PMGM, and C-PDSM are respectively 1, K, and K, where K is the number of solutions.
Proof. In the Calculation phase, the agent in C-MGM sends a random value to its neighbors to calculate the local utility, the message size of the random value is 1. However, the agents in C-PMGM and C-PDSM send value assignments to neighbors. The size of value assignments is K, where K is the number of solutions. In the Comparison phase, the size of the gain message in C-MGM is 1, but the size of gain sets in the other two algorithms is K. Hence, since , , and , the message size of C-MGM, C-PMGM, and C-PDSM are respectively 1, K, and K in one iteration. □
Theorem 4. In one iteration, the computational overhead of one agent in C-MGM, C-PMGM, and C-PSAM are respectively , , and , where K is the number of solutions and is the number of neighbors. Further, the overall computational overhead of C-MGM, C-PMGM, and C-PSAM in one iteration are , , and , respectively.
Proof. We define the number of neighbors of one agent as . For C-MGM, the agent needs to calculate the local utility for K solutions in one iteration, including the basic local utility. In addition, the agent calculates the gain once. Therefore, in one iteration, the computational overhead of one agent in C-MGM is . Since , and the overall computational overhead of C-MGM is .
For C-PMGM, the agent calculates the basic local utility and competing local utility (Algorithm 3: Line 15) for K solutions with neighbors, and the agent calculates the gain set, including K gains. Hence, the computational overhead of one agent in C-PMGM is , the overall computational overhead is .
Compared to C-PMGM, an agent of C-PDSM additionally updates the value of K solutions based on C-PMGM. Therefore, the computational overhead of one agent in C-PDSM in one iteration is , and the overall computational overhead is . □
5. Experimental Results
In this section, we empirically evaluate the solution quality of our three algorithms and the state-of-the-art algorithms, PCD, C-CoCoA, and PCD-LD. Firstly, we set initialization parameters for each algorithm. Specially, we evaluate the performance of C-PDSM with different scaling factors to find the most suitable scaling factor for comparison experiments. Then, we compare the solution quality of the six algorithms when the algorithm runs completely. Finally, we use the runtime to evaluate the performance of all algorithms.
Although C-MGM, C-PMGM, and C-PDSM can use any form of the utility function, we use the utility function in the form of
to fairly show the results for consistency with researchers’ work [
38,
40], where
, and
f are random numbers in the range
. Furthermore, we set the domain of each variable to
and the number of iterations to 500. It is worth mentioning that we choose the best solutions in C-PMGM and C-PDSM to present comparisons. For each benchmark problem, we independently run each algorithm on 30 different sets of function coefficients (
, and
f) and take the average solution quality as the experimental result. The experiments are carried out on a computer equipped with Intel(R) Core(TM) i5-10500 CPU, 3.10GHz processor, and 8GB RAM.
We choose three benchmark problems and set their parameters:
Random Graphs: we use the Erdös-Rényi model [
43] to generate random graphs with two different settings:
D1: Sparse random graphs (
) and
D2: dense random graphs (
), where
p represents the probability of connection between two nodes in a graph;
D3:
Scale-free Networks: we refer to the Barabási-Albert (BA) [
44] model to generate scale-free networks with an initial set of
connected agents, and at each iteration, a new agent that connected to
other agents is added to the current network; and
D4:
Small-world Networks: we use the Newman-Watts model [
45] to provide small-world networks. The number of nearest nodes is set to 8 and the probability of connection adding is set to 0.1.
5.1. Parameter Configuration
In C-MGM, C-PMGM, and C-PDSM,
K represents the size of the solution set, which is also the number of solutions that need to be optimized. Considering the previous work of some researchers using the solution set [
37,
40], we uniformly set
K to 1000. For the scaling factor in C-PDSM, we use different scaling factors to compare the solution quality of C-PDSM on sparse random graphs with 100 agents.
Figure 2 shows the convergence curves of C-PDSM with different scaling factors, where an obvious parameter setting is
. When
, the convergence speed of C-PDSM becomes slower as
decreases. Conversely, the solution quality of C-PDSM with different parameters is similar when
. However, we can observe that there is a slight superiority in terms of convergence when
. Therefore, we set
to 1.6 for C-PDSM. In addition, we set the parameters for competing algorithms according to the references [
37,
39,
40].
5.2. Comparison of Solution Quality
We show convergence curves of our algorithms and competing algorithms on three benchmark problems with 100 agents. Since C-CoCoA is a non-iterative algorithm, we use a straight line to represent its solution quality.
Figure 3a and
Figure 3b show the solution quality of the six algorithms on sparse and dense random graphs, respectively. It can be seen that the solution quality of our three algorithms is significantly better than that of competing algorithms. Compared to performance on sparse random graphs, our three algorithms require more iterations to improve the solution quality. The most critical reason is that each agent has more neighbors in the dense configuration. Since only one of the two neighboring agents is allowed to change its value in our algorithms, there are fewer agents to improve their local utilities in one iteration. Therefore, our algorithms converge slower. Nevertheless, the solution quality of the three algorithms is superior to competing algorithms at about 100 iterations.
The convergence curves of the six algorithms on scale-free and small-world networks are presented in
Figure 3c and
Figure 3d, respectively. The performance of our algorithms is similar to that on sparse random graphs. C-MGM can obtain a high-quality solution and converge after a few iterations, and C-PMGM can improve the solution quality as the number of iterations increases. Compared to C-MGM, C-PDSM can converge faster since it uses heuristics to enhance the exploitation of solutions.
Further, we evaluate the solution quality of the six algorithms on each benchmark problem with 80 and 90 agents, respectively. The detailed results are shown in
Table 1. We can observe clearly that the solution quality of C-MGM, C-PMGM, and C-PDSM significantly outperforms that of competing algorithms on different benchmark problems with varying numbers of agents, and C-PDSM has the best solution quality among all algorithms.
5.3. Comparison of Algorithms Using Runtime
We compare the best solution quality reached by C-MGM, C-PMGM, C-PDSM, and competing algorithms on each benchmark problem after runtime in 1000 milliseconds (ms).
Table 2 lists the solution quality of algorithms on benchmark problems with different numbers of agents. Since C-MGM uses multiple local utilities to find the maximum gain compared to C-PMGM and C-PDSM, it can quickly find a high-quality solution. In addition, the solution quality of C-MGM outperforms competing algorithms on each benchmark problem except for the comparison with C-CoCoA on dense random graphs, the reason is that C-MGM converges slower due to the increase of neighbors.
6. Conclusion and Future Work
Researchers propose C-DCOPs to model distributed problems with continuous variables as an extension of DCOPs. In this paper, we design three algorithms C-MGM, C-PMGM, and C-PDSM for solving C-DCOPs and prove that they are three anytime algorithms. Since the C-MGM algorithm resets value assignments to expend the search space and chooses the maximum gain, it can obtain high-quality solutions with few iterations. The C-PMGM algorithm extends C-MGM by searching multiple solutions in parallel, which can significantly improve the solution quality. Based on C-PMGM, C-PDSM combines differential search to strengthen the exploitation of solutions, which can effectively improve the convergence speed and the solution quality. Extensive experiments on three benchmark problems show that C-MGM, C-PMGM, and C-PDSM outperform the state-of-the-art C-DCOP algorithms. In future work, we plan to improve the convergence speed of the three algorithms on some dense graphs.
Author Contributions
Conceptualization, Xin Liao and Khoi Hoang; methodology, Xin Liao; software, Xin Liao; validation, Xin Liao; formal analysis, Xin Liao; investigation, Xin Liao; resources, Khoi Hoang; data curation, Xin Liao; writing—original draft preparation, Xin Liao; writing—review and editing, Xin Liao; visualization, Khoi Hoang; supervision, Khoi Hoang; project administration, Khoi Hoang; funding acquisition, Khoi Hoang. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
Conflicts of Interest
The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
References
- Modi, P.; Shen, W.M.; Tambe, M.; Yokoo, M. ADOPT: Asynchronous Distributed Constraint Optimization with Quality Guarantees. Artificial Intelligence 2005, 161, 149–180. [CrossRef]
- Petcu, A.; Faltings, B. A Scalable Method for Multiagent Constraint Optimization. In Proceedings of the Proceedings of the 19th International Joint Conference on Artificial Intelligence, 2005, IJCAI, pp. 266–271.
- Yeoh, W.; Yokoo, M. Distributed Problem Solving. AI Magazine 2012, 33, 53–65.
- Fioretto, F.; Pontelli, E.; Yeoh, W. Distributed Constraint Optimization Problems and Applications: A Survey. Journal of Artificial Intelligence Research 2018, 61, 623–698. [CrossRef]
- Maheswaran, R.T.; Tambe, M.; Bowring, E.; Pearce, J.P.; Varakantham, P. Taking DCOP to the Real World: Efficient Complete Solutions for Distributed Multi-Event Scheduling. In Proceedings of the Proceedings of the 3rd International Conference on Autonomous Agents and Multiagent Systems, 2004, AAMAS, pp. 310–317.
- Di, B.; Jennings, N.R. A Contract-based Incentive Mechanism for Distributed Meeting Scheduling: Can Agents Who Value Privacy Tell the Truth? Autonomous Agents and Multiagent Systems 2021, 35, 35.
- Farinelli, A.; Rogers, A.; Jennings, N.R. Agent-Based Decentralised Coordination for Sensor Networks Using the Max-sum Algorithms. Autonomous Agents and Multiagent Systems 2014, 28, 337–380. [CrossRef]
- Zivan, R.; Yedidsion, H.; Okamoto, S.; Glinton, R.; Sycara, K.P. Distributed Constraint Optimization for Teams of Mobile Sensing Agents. Autonomous Agents and Multiagent Systems 2015, 29, 495–536. [CrossRef]
- Harel, Y.; Roie, Z. Applying DCOP_MST to a Team of Mobile Robots with Directional Sensing Abilities. In Proceedings of the Proceedings of the 15th International Conference on Autonomous Agents and Multiagent Systems, 2016, AAMAS, pp. 1357–1358.
- Fioretto, F.; Yeoh, W.; Pontelli, E.; Ma, Y.; Ranade, S.J. A Distributed Constraint Optimization (DCOP) Approach to the Economic Dispatch with Demand Response. In Proceedings of the Proceedings of the 16th International Conference on Autonomous Agents and Multiagent Systems, 2017, AAMAS, pp. 999–1007.
- Miller, S.; Ramchurn, S.D.; Rogers, A. Optimal Decentralised Dispatch of Embedded Generation in the Smart Grid. In Proceedings of the Proceedings of the 11st International Conference on Autonomous Agents and Multiagent Systems, 2012, AAMAS, pp. 281–288.
- Rust, P.; Picard, G.; Ramparany, F. Using Message-Passing DCOP Algorithms to Solve Energy-Efficient Smart Environment Configuration Problems. In Proceedings of the Proceedings of the 25th International Joint Conference on Artificial Intelligence, 2016, IJCAI, pp. 468–474.
- Fioretto, F.; Yeoh, W.; Pontelli, E. A Multiagent System Approach to Scheduling Devices in Smart Homes. In Proceedings of the Proceedings of the 16th International Conference on Autonomous Agents and Multiagent Systems, 2017, AAMAS, pp. 981–989.
- Chen, Z.; Deng, Y.; Wu, T. An Iterative Refined Max-sum_AD Algorithm via Single-side Value Propagation and Local Search. In Proceedings of the Proceedings of the 16th International Conference on Autonomous Agents and Multiagent Systems, 2017, AAMAS, pp. 195–202.
- Chen, Z.; Wu, T.; Deng, Y.; Zhang, C. An Ant-based Algorithm to Solve Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 32nd AAAI Conference on Artificial Intelligence, 2018, AAAI, pp. 4654–4661.
- Hoang, K.D.; Fioretto, F.; Yeoh, W.; Pontelli, E.; Zivan, R. A Large Neighboring Search Schema for Multi-agent Optimization. In Proceedings of the Proceedings of the 24th International Conference on Principles and Practice of Constraint Programming, 2018, CP, pp. 688–706.
- Khan, M.M.; Tran-Thanh, L.; Jennings, N.R. A Generic Domain Pruning Technique for GDL-Based DCOP Algorithms in Cooperative Multi-Agent Systems. In Proceedings of the Proceedings of the 17th International Conference on Autonomous Agents and Multiagent Systems, 2018, AAMAS, pp. 1595–1603.
- Grinshpoun, T.; Tassa, A.; Levit, V.; Zivan, R. Privacy Preserving Region Optimal Algorithms for Symmetric and Asymmetric DCOPs. Artificial Intelligence 2019, 266, 27–50. [CrossRef]
- Tassa, T.; Grinshpoun, T.; Yanai, A. A Privacy Preserving Collusion Secure DCOP Algorithm. In Proceedings of the Proceedings of the 28th International Joint Conference on Artificial Intelligence, 2019, IJCAI, pp. 4774–4780.
- Tassa, T.; Grinshpoun, T.; Yanai, A. PC-SyncBB: A Privacy Preserving Collusion Secure DCOP Algorithm. Artificial Intelligence 2021, 297, 103501.
- Deng, Y.; Chen, Z.; Chen, D.; Zhang, W.; Jiang, X. AsymDPOP: Complete Inference for Asymmetric Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 28th International Joint Conference on Artificial Intelligence, 2019, IJCAI, pp. 223–230.
- van Leeuwen, C.J.; Pawelczak, P. CoCoA: A Non-Iterative Approach to a Local Search (A)DCOP Solver. In Proceedings of the Proceedings of the 35th AAAI Conference on Artificial Intelligence, 2017, AAAI, pp. 3944–3950.
- Zivan, R.; Parash, T.; Cohen-Lavi, L.; Naveh, Y. Applying Max-sum to Asymmetric Distributed Constraint Optimization Problems. Autonomous Agents and Multiagent Systems 2020, 34, 1–29. [CrossRef]
- Deng, Y.; Chen, Z.; Chen, D.; Jiang, X.; Li, Q. PT-ISABB: A Hybrid Tree-based Complete Algorithm to Solve Asymmetric Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 18th International Conference on Autonomous Agents and Multiagent Systems, 2019, AAMAS, pp. 1506–1514.
- Hoang, K.D.; Hou, P.; Fioretto, F.; Yeoh, W.; Zivan, R.; Yokoo, M. Infinite-Horizon Proactive Dynamic DCOPs. In Proceedings of the Proceedings of the 16th International Conference on Autonomous Agents and Multiagent Systems, 2017, AAMAS, pp. 212–220.
- Hoang, K.D.; Fioretto, F.; Hou, P.; Yokoo, M.; Yeoh, W.; Zivan, R. Proactive Dynamic Distributed Constraint Optimization. In Proceedings of the Proceedings of the 15th International Conference on Autonomous Agents and Multiagent Systems, 2016, AAMAS, pp. 597–605.
- Hoang, K.D.; Fioretto, F.; Hou, P.; Yeoh, W.; Yokoo, M.; Zivan, R. Proactive Dynamic Distributed Constraint Optimization Problems. Journal Of Artificial Intelligence Research 2022, 74, 179–225. [CrossRef]
- Gutierrez, P.; Meseguer, P.; Yeoh, W. Generalizing ADOPT and BnB-ADOPT. In Proceedings of the Proceedings of the 22th International Joint Conference on Artificial Intelligence, 2011, IJCAI, pp. 554–559.
- Yeoh, W.; Felner, A.; Koenig, S. BnB-ADOPT: An Asynchronous Branch-and-Bound DCOP Algorithm. Journal of Artificial Intelligence Research 2010, 38, 85–133.
- Vinyals, M.; Rodríguez-Aguilar, J.A.; Cerquides, J. Constructing A Unifying Theory of Dynamic Programming DCOP Algorithms via the Generalized Distributive Law. Autonomous Agents and Multiagent Systems 2011, 22, 439–464. [CrossRef]
- Maheswaran, R.T.; Pearce, J.P.; Tambe, M. Distributed Algorithms for DCOP: A Graphical-Game-Based Approach. In Proceedings of the Proceedings of the 17th International Conference on Parallel and Distributed Computing Systems, 2004, PDCS, pp. 432–439.
- Ottens, B.; Dimitrakakis, C.; Faltings, B. DUCT: An Upper Confidence Bound Approach to Distributed Constraint Optimization Problems. ACM Transactions on Intelligent Systems and Technology 2017, 8, 69:1–69:27.
- Nguyen, D.T.; Yeoh, W.; Lau, H.C.; Zivan, R. Distributed Gibbs: A Linear-Space Sampling-Based DCOP Algorithm. Journal of Artificial Intelligence Research 2019, 64, 705–748. [CrossRef]
- Stranders, R.; Farinelli, A.; Rogers, A.; Jennings, N.R. Decentralised Coordination of Continuously Valued Control Parameters Using the Max-sum Algorithm. In Proceedings of the Proceedings of the 8th International Conference on Autonomous Agents and Multiagent Systems, 2009, AAMAS, pp. 601–608.
- Voice, T.; Stranders, R.; Rogers, A.; Jennings, N.R. A Hybrid Continuous Max-Sum Algorithm for Decentralised Coordination. In Proceedings of the Proceedings of the 19th European Conference on Artificial Intelligence, 2010, ECAI, pp. 61–66.
- Fransman, J.; Sijs, J.; Dol, H.; Theunissen, E.; Schutter, B.D. Bayesian-DPOP for Continuous Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 18th International Conference on Autonomous Agents and Multiagent Systems, 2019, AAMAS, pp. 1961–1963.
- Choudhury, M.; Mahmud, S.; Khan, M.M. A Particle Swarm Based Algorithm for Functional Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 34th AAAI Conference on Artificial Intelligence, 2020, AAAI, pp. 7111–7118.
- Hoang, K.D.; Yeoh, W.; Yokoo, M.; Rabinovich, Z. New Algorithms for Continuous Distributed Constraint Optimization Problems. In Proceedings of the Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems, 2020, AAMAS, pp. 502–510.
- Sarker, A.; Choudhury, M.; Khan, M.M. A Local Search Based Approach to Solve Continuous DCOPs. In Proceedings of the Proceedings of the 20th International Conference on Autonomous Agents and Multiagent Systems, 2021, AAMAS, pp. 1127–1135.
- Shi, M.; Liao, X.; Chen, Y. A Particle Swarm with Local Decision Algorithm for Functional Distributed Constraint Optimization Problems. International Journal of Pattern Recognition and Artificial Intelligence 2022, 36, 2259025. [CrossRef]
- Zivan, R.; Okamoto, S.; Peled, H. Explorative Anytime Local Search for Distributed Constraint Optimization. Artificial Intelligence 2014, 212, 1–26. [CrossRef]
- Storn, R.; Price, K.V. Differential Evolution - A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces. Journal of Global Optimization 1997, 11, 341–359. [CrossRef]
- Paul, E.; Rényi, A. On the Evolution of Random Graphs. Publications of the Mathematical Institute of the Hungarian Academy of Sciences 1960, 5, 17–60.
- Albert-László, B.; Réka, A. Emergence of Scaling in Random Networks. Science 1999, 286, 509–512.
- Newman, M.; Watts, D. Renormalization Group Analysis of the Small-World Network Model. Physics Letters A 1999, 263, 341–346. [CrossRef]
1 |
We are going to consider the minimization in this paper. |
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).