Published online by Cambridge University Press: 12 February 2004
In this paper we are interested in the decision problem faced by an agent when requesting bids for collections of tasks with complex time constraints and interdependencies. In particular, we study the problem of specifying an appropriate schedule for the tasks in the request for bids. We expect bids to require resource commitments, so we expect different settings of time windows to solicit different bids and different costs. The agent is interested in soliciting “desirable” bids, where desirable means bids that can be feasibly combined in a low-cost combination that covers the entire collection of tasks. Since the request for bids has to be issued before the agent can see any bids, in this decision process there is a probability of loss as well as a probability of gain. This requires the decision process to deal with the risk posture of the person or organization on whose behalf the agent is acting. We describe a model based on Expected Utility Theory and show how an agent can attempt to maximize its profits while managing its financial risk exposure. We illustrate the operation and properties of the model and discuss what assumptions are required for its successful integration in multiagent contracting applications.
E-commerce technology has the potential to benefit society by reducing the cost of buying and selling and by opening new market opportunities. We envision an auction-based approach to the management of agile and dynamic supply chains, in which autonomous, self-interested agents negotiate on behalf of organizations and individuals to organize coordinated activities. This is an area in which the potential payoff is very high, given the projected size of the business to business and make to order e-commerce markets.
More production processes are being outsourced to outside contractors, making supply chains longer and more convoluted. This increased complexity is compounded by increasing competitive pressure and accelerated production schedules that demand tight integration of all processes. Finding potential suppliers is only one step in the process of producing goods. Time dependencies among operations make scheduling a major factor. A late delivery of a part may produce a cascade of devastating effects. Unfortunately, current auction-based systems do not have any notion of time. Handling auctions for tasks with time constraints is beyond the capabilities of current e-commerce systems.
We present the results of a study of how an autonomous agent can maximize its profits while predicting and managing its financial risk exposure when requesting bids for tasks with complex time constraints. We show how this can be done by specifying appropriate time windows for tasks when soliciting bids and by using received bids effectively in building a final work schedule.
This study is a part of the MAGNET (multiagent negotiation testbed) research project (Collins et al., 2002). MAGNET agents participate in first-price, sealed-bid combinatorial auctions over collections of tasks with precedence relations and time constraints. MAGNET promises to increase the efficiency of current markets by shifting much of the burden of market exploration, auction handling, and preliminary decision analysis from human decision makers to a network of heterogeneous agents.
We distinguish between two agent roles, the Customer and the Supplier (see Fig. 1). A customer has a set of tasks to be performed, with complex dependencies among the tasks. When a customer lacks the resources to carry out its own tasks, it may solicit resources from suppliers by presenting a request for quotes (RFQ) through an agent-mediated market. Supplier agents may offer to provide the requested resources or services, for specified prices, over specified time periods. Once the customer receives bids, it evaluates them to select an optimal set of bids and create a work schedule. This paper deals with decision problems in the Bid Manager component of the Customer Agent.
The MAGNET architecture.
This is a schematic outline of the main interactions among agents:
As an example, imagine that we need to construct a garage. Figure 2 shows the tasks needed to complete the construction. The tasks are represented in a task network, where links indicate precedence constraints. The first decision we are faced with is how to sequence the tasks in the RFQ and how much time to allocate to each of them. For instance, we could reduce the number of parallel tasks, allocate more time to tasks with higher variability in duration or for which there is a shortage of laborers, or allow more slack time.
A task network example and a corresponding RFQ.
A sample RFQ is shown in Figure 2. Note that the time windows in the RFQ do not need to obey the precedence constraints; the only requirement is that the accepted bids obey them. We assume that the supplier is more likely to bid, and submit a lower-cost bid, if it is given a greater flexibility in scheduling its resources. It is up to the customer to find a bid combination that forms a feasible schedule.
We have shown (Collins et al., 2001) that the time constraints specified in the RFQ can affect the customer's outcome in two major ways:
The agent needs to issue the RFQ before having received any bid, so the process of deciding how to schedule the different tasks and how much time to allocate to each task involves a probability of loss, as well as a probability of gain. This requires the decision process to deal with the risk posture of the person or organization on whose behalf the agent is acting.
The agent can use information available from the market on expected costs, probability of completion of tasks within a time window, and expected numbers of bidders to guide its decision on how to sequence the tasks and how much time to allocate to each.
In the next section we propose a principled method for generating RFQs that takes into account the agent's risk posture and available market statistics to produce a schedule that optimizes the agent's expected utility.
In this section we describe a new approach to the construction of optimal RFQs that employs the Expected Utility Theory to reduce the likelihood of receiving unattractive bids while maximizing the number of bids that are likely to be awarded. This approach was originally suggested in our previous work (Babanov, Collins, et al., 2002). In this work we extend it and pay special attention to the relation between the size of RFQ time windows and the number of expected bids by investigating the balance between the quantity and the quality of expected bids.
A task network (see Fig. 2) is a tuple 〈N, <〉 of a set N of individual tasks and strict partial ordering on them, such that for any i, j ∈ N, i < j implies that task i immediately precedes task j. We also use N to denote the number of tasks where appropriate.
A task network is characterized by a start time (ts) and a finish time (tf), which delimit the interval of time when tasks can be scheduled. The placement of task n in the schedule is characterized by task n start time (tns) and task n finish time (tnf), subject to the following constraints:
where P1(n) is the a set of immediate predecessors of n, P1(n) = {m ∈ N|m < n}, and S1(n) is defined similarly to be the set of immediate successors of task n.
The probability of task n completion by time t, conditional on the ultimate successful completion of task n, is distributed according to the cumulative distribution function (CDF) Φn = Φn(tns; t), limt→∞ Φn(tns; t) = 1. Observe that Φn is defined to be explicitly dependent on the tns. To see the rationale, consider the probability of successful mail delivery in x days for packages that were mailed on different days of a week.
There is an associated unconditional probability of success pn ∈ [0, 1] characterizing the percentage of tasks that are successfully completed given infinite time (see Fig. 3). In the empirical support of work we assumed Weibull probability distribution for Φn; however, the form of the distribution is not tied in the theory. In fact, we expect that the success probabilities will be derived from the available market information.
The unconditional distribution for the successful completion probability.
Task n bears an associated cost.1
Hereafter we use the words “cost” and “reward” to denote some monetary value, while referring the same value as “payment” or “payoff” whenever it is scheduled at some time t.
There is a single final reward V scheduled at the plan finish time tf and paid conditional on all tasks in N being successfully completed by that time.
For each cost and reward, there is an associated rate of return2
The reason for having multiple qn values is that individual tasks can be financed from different sources, thus affecting task scheduling.
We associate the rate of return q with the final payoff V.
We represent the customer agent's preferences over payoffs by the von Neumann–Morgenstern utility function u (Mas–Colell et al., 1995). We further assume that the absolute risk-aversion coefficient r := −u′′/u′ of u is constant for any value of its argument; hence, u can be represented as follows:
It is imperative to note here that we do not compare utility values directly; the counterintuitive (i.e., decreasing in monetary terms) form of the utility for r < 0 is a trade-off for simple notation.
We assume that a future state of the world can be described in terms of potential money transfers and the corresponding probabilities. Accordingly, we define gamble to be a set of payoff–probability pairs G = {(xi, pi)i} s.t. pi > 0, ∀i and [sum ]i pi = 1. The expectation of the utility function over a gamble G is the expected utility (EU):
The certainty equivalent (CE) of a gamble G is defined as the single payoff value whose utility matches the expected utility of the entire gamble G, that is, u(CE[G]) := EU[G] . Hence, under our assumptions
Our evaluation criterion is based upon comparing CE values, because they represent money transfers in certain and current money. Because of this interpretation the CE values, unlike utilities, can be compared across various risk aversities and alternative schedules, even between different plans. Naturally, the agent will not be willing to accept gambles with a negative certainty equivalent and higher values of the certainty equivalent will correspond to more attractive gambles.
To illustrate the concept, Figure 4 shows how the CE depends on the risk aversity r of an agent. In this figure we consider a gamble that brings the agent either 100 or nothing with equal probabilities. Agents with positive r values are risk averse; those with negative r values are risk loving. Agents with risk aversity close to zero, that is, almost risk neutral, have a CE equal to the gamble's weighted mean (50).
The CE of a simple gamble as a function of the risk aversity.
To compute the certainty equivalent of a gamble we need to determine a schedule for the tasks and compute the payoff probability pairs.
We assume that the cn for task n is scheduled at tnf, so its present value
3Hereafter we use the tilde to distinguish variables that depend on the current task schedule, while omitting corresponding indices for the sake of notational simplicity.
is
We define the conditional probability of task n success as
We also define the precursors of task n as a set of tasks that finish before task n starts in a schedule, that is,
The unconditional probability that task n will be completed successfully is
That is, the probability of successful completion of every precursor and of task n itself are considered as independent events. The reason this is calculated in such form is because, if any task in
fails to be completed, there is no need to execute task n.
The probability of receiving the final reward V is therefore
To illustrate the definitions above, let us return to the task network in Figure 2 and consider the sample task schedules shown in Figure 5. In this figure the x axis is time and the y axis shows both the task numbers and the cumulative distribution of the unconditional probability of completion (compare to Fig. 3). Circle markers show tns. Crosses indicate both tnf and success probabilities
(numbers next to each point). Square markers denote that the corresponding task cannot span past this point due to precedence constraints. The thick part of each CDF shows the time allocated to each task.
The CE maximizing time allocations for the plan in Figure 2 for r = −0.01 (left) and r = 0.02 (right).
The customer agent needs a way of collecting the market information necessary to build and use the probability model. The probability of success is relatively easy to observe in the market. This is the reason for introducing the cumulative probability of success Φn and probability of success pn, instead of the average project life span or probability of failure. Indeed, it is rational for the supplier to report a successful completion immediately in order to maximize the present value of a payment. In addition, it is rational not to report a failure until the last possible moment, because of a possibility of earning the payment by rescheduling, outsourcing, or fixing the problem in some way.
Given a schedule like the one shown in Figure 5, we need to compute the payoff probability and then maximize the CE for the gamble. Writing an explicit description of the expected utility as a function of gambles is overly complicated and relies on the order of task completions. Instead, we propose a simple recursive algorithm that creates these gambles. We then maximize the CE over the space of all feasible schedules and the corresponding gambles.
In the first call, the algorithm receives a “todo” task list T = N and a “done” task list D = [empty ]. All the subsequent calls are recursive. To illustrate the idea behind this algorithm, we refer to the payoff–probability trees in Figure 6. These two trees were built for the time allocations in Figure 5 and reflect the precursor relations for each case.
Payoff–probability trees for the time allocations in Figure 5.
Considering the more sequential schedule on the right, we note that with probability
ask 1 fails, and the customer agent does not pay or receive anything and stops the execution (path 1 in the right tree). With probability
the agent proceeds with task 3 (path 1 in the tree). In turn, task 3 either fails with probability
, in which case the agent ends up stopping the plan and paying a total of c1 (path 1 → 3), or it is completed with the corresponding probability
. In the case where both 1 and 3 are completed, the agent starts both 2 and 4 in parallel and becomes liable for paying c2 and c4, respectively, even if the other task fails (paths 1 → 3 → 2 → 4 and 1 → 3 → 2 → 4). If both 2 and 4 fail, the resulting path in the tree is 1 → 3 → 2 → 4 and the corresponding payoff–probability pair is framed in the figure.
The algorithm's complexity is O(2K−1 × N), where K is the maximum number of tasks that are scheduled to be executed in parallel. Reducing the complexity of calcGamble is critical, because it will be executed in the inner loop of any CE maximization procedure, unless we somehow fix precursor relations and, consequently, a tree structure. In commercial projects the ratio K/N is likely to be low, because few of these exhibit a high degree of parallelism. Our preliminary experiments allow us to conclude that the K/N ratio is lower for risk-averse agents (presumably, businesspeople) than for risk lovers (gamblers). These two considerations may reduce the need for a faster algorithm, although additional work to improve the algorithm is planned.
We have conducted a set of experiments on CE maximization in a variety of task networks. Some of the results for our reference six-task network are summarized in Figure 7. In this figure, the y axis shows 11 different risk-aversity r settings, and the bottom x axis is time t in the plan. The rounded horizontal bars in each of 11 sections denote time allocations for each of six tasks, with task 1 on top. Sections r = −0.01 and r = 0.02 correspond to schedules in Figure 5a and b, respectively. Finally, the vertical bars show the maximum CE values on the top x axis for each r setting.
The CE maximizing schedules and CE values for the plan in Figure 2 and r ∈ [−0.03, 0.07] .
Let us examine the relative placement of time allocations as a function of r. For this example we chose two tasks, 3 and 4, which have similar positions in the task network. Task 3 (black horizontal bars) has a lower probability of success in the infinite horizon than task 4 (white bars), as well as a higher variance of the probability of success distribution. In addition, the cost of task 3 is twice the cost of task 4.
Given this setup, we observed four distinct cases in the experimental data:
In the previous sections we have shown a way of generating a CE maximizing schedule of task execution, which we hereafter refer to as the reasonable schedule. For a chosen risk-aversity value and known market statistics, the reasonable schedule ensures the highest expected quality of the bids that satisfy it. By quality, we assume some function of the expectations over the cost, the probability of successful completion, and the profitability of the incoming bids in their feasible combinations with other bids.
The reasonable schedule, however, cannot serve as a rational RFQ, because it is unlikely that bids will be available to cover precisely the same intervals as mandated by the CE maximizing schedule. In order to construct a viable RFQ using the reasonable schedule as a basis, the customer agent might choose to lower its expectations of the bid quality to some level by widening the RFQ time windows around the time windows in the reasonable schedule, thus increasing4
At least to some extent, because there is a fair chance that the number of the incoming bids will cease to increase whenever RFQ time windows become too large to inspire confidence on the part of suppliers.
We approach the individually rational (i.e., agent-dependent) RFQ generation as follows:
In addition, we search for a solution concept that generates viable RFQs and is comprehensible to a human user of the system.
We propose measuring the sensitivity of CE by investigating how CE values change with variations of a single task n start time (tns) in the reasonable schedule. For the sake of brevity the resulting dependency of CE values is denoted by CE(tns). Figure 8 shows CE(tns), n = 1···6 for our six-task sample problem for r = −0.01 and 0.02, respectively. In the figure, the y axis of each horizontal stripe n represents the percentage of the maximum CE value as tns varies, the x axis represents tns, and the horizontal lines with circle and cross ends show the corresponding reasonable schedules.
The CE(tns) graphs for the corresponding reasonable schedules in Figure 5.
Tasks 1, 3, and 5 in the right graph are relatively restrictive to the start times of the bids that can be bundled with the reasonable bids without considerably impairing the resulting bundle's value. However, the fact that task 2 in the right graph is more flexible does not guarantee that it will attract a higher number of bids, because the latter depend both on the size of the corresponding time window and on the market properties of the task: resource availability, number of prospective bidders, seasonal changes, and so forth.
We assert that for the purpose of creating a rational RFQ, it is admissible to choose time windows based on the sensitivity of CE to deviations of a single time constraint from the reasonable schedule. The rationale is that the relations between tasks are already encapsulated in the calculations of CE, so the change of one constraint will approximate the rescheduling of several related tasks in the neighborhood of the reasonable schedule.
Observe that the time window for the task n, {tns|CE(tns) ≥ x}, grows as the lowest expected CE value (x) decreases. The relation between these two variables for the tasks 3 and 4 of the test problem is shown in Figure 9. The corresponding relation between the lowest expected CE value and the expected number of bids as a function of the window size is shown in Figure 10. In the right graph we assumed, for the sake of example, that the supply of task 3 in the market is higher than task 4, hence the difference in relative positions of task 3 and task 4 graphs in the two figures.
The relationship between the RFQ window size (units of time on x axis) and the lowest admissible percentage of the maximum CE value.
The relationship between the expected number of bids (x axis) and the lowest admissible percentage of the maximum CE value.
The graph in Figure 10 is an example of the relation between the quality and the quantity of bids for which we are searching. Indeed, the only independent variable in this graph is tns. The quantity of bids depends on the size and positions of RFQ time windows that, in turn, depend on the decision about the lowest acceptable CE value. The quality of bids is a function of the RFQ choice and the properties of the plan. Finally, it is expected that the customer agent will prefer a point on the graph to any point below and to the left of it; thus, the best choice should lie on the graph.
We illustrate the decision process of the customer agent in Figure 11, where the customer agent's preferences over quality–quantity combinations are represented by a family of indifference curves and the graph of underlying quality–quantity relationship is as derived in the previous section. Each indifference curve shows quality–quantity pairs that are equivalent from the agent's point of view. In particular, points A, B, and C are considered to be equally attractive. The intuition is that, although in point A the agent receives a much smaller number of bids to compare with C and is exposed to a higher risk of not covering some tasks, this is offset by a positive effect of a lower percentage of bid rejections on the agent's reputation. In addition, the winner determination problem is exponentially easier to solve (Collins, 2002) for the lower expected number of bids in point A.
A quality–quantity graph with three (—) indifference curves; (– – –) the maximum expected quality line.
For all points below the maximum expected quality line (Fig. 11, dashed line) agent's preferences increase in the direction of point M. Thus, a curve through point D is preferred over one through point C and even more so over one through point E. The rational choice belongs to the intersection of the quality–quantity graph and the highest indifference curve (point B).
After the rational choice of the quality–quantity combinations for all tasks in the plan is revealed, we proceed with constructing the RFQ time windows. The choice of early start time (tnes) and late start time (tnls) are determined by the value of the reciprocal of the CE(tns) at the minimum admissible CE choice for task n. The late finish time (tnlf) is chosen to be at the reasonable time window length distance from tnls. Figure 12 shows two sample RFQs for the garage building example. The start time intervals ([tnes, tnls]), late finish time tnlf, and the corresponding reasonable schedules are presented in the figure.
The rational RFQs for the corresponding reasonable schedules in Figure 5 and the maximum CE percentages of 80, 95, 50, 70, 50, and 90%. (shaded bar segments) start time intervals, (unshaded bar segments) late finish times, and (○—×) corresponding reasonable schedules.
Our choice of the RFQ may not be optimal in the quantitative sense. However, it is individually rational for the customer agent, fast to compute, and, arguably, easily grasped by a human user of the system. It should be emphasized here that the choice of the RFQ is based on the uncertain market information; hence, any quantitatively “optimal” solution is itself a compromise.
In this section we outline two major issues that arise when we employ the expected utility approach to generate rational RFQs. The first issue concerns the CE maximization in the domain with temporal and precedence constraints. The second issue is the assessment of the EU approach and, ultimately, the MAGNET system itself in the absence of the real-world data for the domain of interest.
One of the most important issues related to CE maximization is the presence of multiple local maxima of CE, even in cases where task networks are fairly simple. We argue that this property is partially due to the relative positioning of the tasks off the critical path. Any two tasks that are not ordered by the precedence constraints can be scheduled in three ways: parallel and two sequential. Scheduling tasks in parallel increases the probability of successful completion, whereas sequential scheduling minimizes overall payments in case one of the tasks fails. In cases where extra slack allows for sequential scheduling, it turns out that parallel and sequential positionings of two independent individual tasks lead to similar resulting CE values.
To illustrate the issue, we constructed a sample task network with two parallel tasks. Task 1 has a higher variance of completion time probability and lower probability of success than task 2; everything else is the same. The resulting graph of CE is shown in Figure 13. There are three local maxima with positive CE values in this figure: one that corresponds to task 2 being scheduled first in sequential order, another on the right side corresponding to task 1 being first, and yet another representing both tasks being scheduled at time 0 and executed in parallel. The number of local maxima grows considerably with the number of the tasks that are not restricted by the precedence relationship.
The local maxima for two parallel tasks.
The following list shows the properties of the domain that influence the search algorithm design:
The properties of the domain frame the properties of the search algorithm that we design to fit this domain. Namely, the search algorithm must be able to test different orderings of tasks, should know how to explore groups of similar local maxima, and, whenever possible, should provide alternative schedules with CE values close to the global maximum.
We propose a search algorithm based on the ideas of simulated annealing (Reeves, 1993) and genetic algorithms (Forrest, 1993). The algorithm will combine the stochastic, temperature-driven nature of simulated annealing with the simultaneous search space exploration of genetic algorithms. In this section we describe the proposed algorithm in more detail and explain the rationale of its design.
The proposed search algorithm explores several alternative schedules in parallel. The initial set of alternatives can be generated in many ways: random generation, hill climbing from a random schedule, critical path method, and so forth. The execution of the algorithm proceeds in steps by randomly applying one of the six transformation rules to each alternative schedule. The probabilities of choosing some rule can be adjusted to adjust the algorithm's behavior over a wide range.
Figure 14 illustrates the algorithm for the case of three pairwise independent tasks. In this figure the columns represent three consecutive states of the algorithm, and each column lists several alternative schedules. Arrows and letters next to the columns denote various transformations from the following list:
The two steps of the search algorithm execution.
Distortion alters start and finish times of one or several tasks and adjusts the time windows of all related tasks to maintain precedence constraints (1 → 5, Fig. 14). Distortion mimics the basic step of the simulated annealing algorithm in continuous 2N–dimensional space of task start and finish times.
Gradient following is one or more steps of a generic numerical maximization method (5 → 9). This step is very computationally intensive, because it requires many calls to the calcGamble algorithm in the process of calculating numerical derivatives. By varying the relative probabilities of the distortion and gradient following, we may choose the stochastic properties of the proposed approach.
Shuffling changes relative scheduling of two or more tasks wherever it is permitted by the precedence constraints. It can switch ordering of the tasks (6 → 10), change the sequential ordering to parallel, or reschedule parallel tasks to be executed sequentially (4 → 8). The major role of shuffling is to explore local maxima that have similar CE values due to different scheduling of tasks off the critical path.
Explosion adds a copy of the subject schedule to the list of alternatives (2 → 6, 7). Explosion compliments shuffling by allowing for simultaneous exploration of the groups of similar schedules. We may choose to decrease the rate of explosions with the annealing temperature to focus on improving the current set of solutions after the search space has been explored to some extent.
Implosion merges two similar5
Similarity is a function of the distance between two schedules, as between two points in the 2N-dimensional time space.
Removal eliminates alternatives that do not score well relative to others (3 → [empty ]). This transformation takes care of the schedules that are stuck in local maxima with low CE values. The rate of removals grows as the annealing temperature decreases.
Each of the first five transformations is tested against the simulated annealing temperature rule whenever it leads to a decrease in the CE value. In case it is discarded, other transformations are chosen at random and applied until one of them increases the CE or passes the temperature rule.
The probabilities of transformations and the details of the proposed search algorithm's properties are subject to further research. However, it is reasonable to believe that the comprehensive study of the RFQ generation mechanism is only possible in the dynamic market environment. In the next section we outline the approach to the large-scale testing of the MAGNET system that we are presently researching and that will provide us with the necessary data.
We plan to devote efforts to thoroughly test the suggested CE-based approach of rational RFQ generation. In particular, we are interested in testing how well individual agents interact in a populated market. The major goals of this part of the study would be the following:
The most compelling approach would be to gather a rich set of statistical data from a commerce domain. That has not proven to be feasible for two reasons. First, few industrial organizations are sufficiently open to expose the type of data you would need to do that, and we would need data from multiple organizations in a single market. Second, data is gathered to serve a purpose, and our experience tells us that when you attempt to apply existing data to a new purpose, it frequently turns out to be full of inconsistencies and methodological problems.
In lieu of using real industry data, we are designing a large-scale test suite atop an abstract domain with controllable statistics, based on the evolutionary approach to economic simulation. Evolutionary frameworks have been used extensively in economics (Nelson, 1995; Rode, 1997; Tesfatsion, 2001). The framework will allow us to tune the market by tweaking the frequency of issuing RFQs and will allow for the dynamic introduction of new supplier strategies without imposing any assumptions on the nature of strategies. We will later extend the framework to support trade games played by human subjects. This will be an especially useful tool for teaching and a tool to explore strategic behaviors and study the emergence of cooperation (Axelrod, 1984; Axelrod, 1997).
We suggest detailed reasoning behind our choice of the evolutionary approach and describe experimental results from a pilot model of a trading agent society in our related research (Babanov, Ketter, et al., 2002).
Expected Utility Theory (Pratt, 1964) is a mature field of economics that has attracted many supportive and critical studies, both theoretical (Machina, 1987, 1989) and empirical (Smith & Desvousges, 1987; Jullien & Salanié, 2000). We believe that expected utility will play an increasing role in automated auctions, because it provides a practical way of describing risk estimations and temporal preferences.
Despite the abundance of work in auctions (McAfee & McMillan, 1987), limited attention has been devoted to auctions over tasks with complex time constraints and interdependencies. Parkes and Ungar (2001) propose a method to auction a shared track line for train scheduling. The problem is formulated with mixed integer programming with many domain-specific optimizations. Bids are expressed by specifying a price to enter a line and a time window. Time slots are used in Wellman et al. (2001), where a protocol for decentralized scheduling is proposed. The study is limited to scheduling a single resource. MAGNET agents deal with multiple resources.
Most work in supply-chain management is limited to hierarchical modeling of the decision making process, which is inadequate for distributed supply chains, where each organization is self-interested rather than cooperative. Walsh et al. (2000) propose a protocol for combinatorial auctions for supply chain formation, using a game-theoretical perspective. They allow complex task networks but do not include time constraints. MAGNET agents must also ensure the scheduling feasibility of the bids they accept and evaluate risk. Agents in MASCOT (Sadeh et al., 1999) coordinate scheduling with the user, but there is no explicit notion of payments or contracts and no explicit statement of the criteria for accepting/rejecting a bid. Their major objective is to show policies that optimize schedules locally (Kjenstad, 1998), whereas our objective is to optimize the customer's utility.
Different heuristics for scheduling are proposed in Dutta et al. (2001). The strategies are intended fo supplier agents that are trying to adjust their schedules to win new awards. In the work presented here, we are concerned with the scheduling that customers need to do before requesting bids.
In MAGNET the agents interact with each other through a market. The market infrastructure provides a common vocabulary, collects statistical information that helps agents estimate costs, schedules, and risks, and acts as a trusted intermediary during the negotiation process. The market also acts as a matchmaker (Sycara et al., 1997), allowing us to ignore the issue of how agents will find each other. For a survey on the use of intelligent agents in manufacturing, see Shen and Norrie (1999).
The determination of winners of combinatorial auctions (Rothkopf et al., 1998) is difficult. Dynamic programming (Rothkopf et al., 1998) works well for small sets of bids, but it does not scale and imposes significant restrictions on the bids. Algorithms such as CABOB (Sandholm et al., 2001), Bidtree (Sandholm, 2000), and CASS (Fujishima et al., 1999) reduce the search complexity. Reeves et al. (2001) use auction mechanisms to “fill in the blanks” in prototype declarative contracts that are specified in a language based on Courteous Logic Programming (Grosof et al., 1999). These auctions support bidding on many attributes other than price, but the problem of combining combinatorial bids with side constraints is not addressed.
Combinatorial auctions are becoming an important mechanism not only for agent-mediated e-commerce (Guttman et al., 1998; Wurman et al., 1998; Sandholm, 1999) but also for the allocation of tasks to cooperative agents (e.g., Dias & Stentz, 2000; Hunsberger & Grosz, 2000).
Hunsberger and Grosz (2000) combinatorial auctions for the initial commitment decision problem, which is the problem an agent must solve when deciding whether to join a proposed collaboration. Their agents have precedence and difficult temporal constraints. However, to reduce search effort, they use domain-specific roles, a shorthand notation for collections of tasks. In their formulation, each task type can be associated with only a single role. MAGNET agents are self-interested, and there are no limits to the types of tasks they can decide to do. In Glass and Grosz (2000) scheduling decisions are made, not by the agents, but instead by a central authority. The central authority has insight to the states and schedules of participating agents, and agents rely on the authority for supporting their decisions.
Leyton–Brown et al. (2000) suggest a way of constructing a universal test suite for winner determination algorithms in combinatorial auctions. Their work does not include cases with precedence and time constraints and, thus, is not directly applicable to the MAGNET framework. It nevertheless provides well-understood test cases for comparing the performance of algorithms.
Auction mechanisms are an effective approach to negotiation among groups of self-interested economic agents. We are particularly interested in situations where agents need to negotiate over multiple factors, including not only price, but task combinations and temporal factors as well.
We have shown how an agent can use information about the risk posture of its principal, along with market statistics, to formulate RFQs that optimize the trade-off between risk and value and increase the quality of the bids received. This requires deciding how to sequence tasks and how much time to allocate to each of them. Bids closest to the specified time windows are the most preferred risk–payoff combinations.
The work described here is a part of a larger effort at the University of Minnesota that aims to study how autonomous or semiautonomous agents can be used in complex commerce-oriented domains.
We gratefully acknowledge the partial support of this research by the National Science Foundation (NSF/IIS-0084202).
The MAGNET architecture.
A task network example and a corresponding RFQ.
The unconditional distribution for the successful completion probability.
The CE of a simple gamble as a function of the risk aversity.
The CE maximizing time allocations for the plan in Figure 2 for r = −0.01 (left) and r = 0.02 (right).
Payoff–probability trees for the time allocations in Figure 5.
The CE maximizing schedules and CE values for the plan in Figure 2 and r ∈ [−0.03, 0.07] .
The CE(tns) graphs for the corresponding reasonable schedules in Figure 5.
The relationship between the RFQ window size (units of time on x axis) and the lowest admissible percentage of the maximum CE value.
The relationship between the expected number of bids (x axis) and the lowest admissible percentage of the maximum CE value.
A quality–quantity graph with three (—) indifference curves; (– – –) the maximum expected quality line.
The rational RFQs for the corresponding reasonable schedules in Figure 5 and the maximum CE percentages of 80, 95, 50, 70, 50, and 90%. (shaded bar segments) start time intervals, (unshaded bar segments) late finish times, and (○—×) corresponding reasonable schedules.
The local maxima for two parallel tasks.
The two steps of the search algorithm execution.