|
|
|
Flame
Towards Flexible Agent Mechanism for the Dynamic Environment
|
 |
|
member |
Hiromitsu HATTORI,Okina YAMAGUTI,Takahumi YAMAYA,Yujiro
HUKAGAYA |
 |
|
keyword |
Multiagent Systems, Negotiation/Argumentation, Constraint Satisfaction
Problems, Soar, Auction, Probabilistic Reasoning, Reinforcement Learning
|
 |
|
mission |
Research project "Flame" elaborates to enhance Multi-agent Systems by constructing next generation of agent technologies. Especially, we try to realize an ability to flexibly handle the dynamics of environment.
|
 |
|
outline |
The member of Flame try to investigate and develop the
sophisticated agent negotiation/coalition formation methods
for deeepening of Multi-agent Systems. In the dynamic
environment, which includes vast amounts of uncertain
information (e.g., Internet), agents must flexibly reason,
negotiate, and make a decision based on the uncertain
information, and they must cosider with the dynamics
of the environment. Therefore, we have to develop/propose
the effective methods for constructing next generation of
agent technologies. Currently, we focus the following subjects.
- The robust problem solving method for the dynamic problem
- The solution for dealing with the solution stability in Dynamic Constraint Satisfaction Problem
- The search/calculation algorithm for effectively adaptation to the dynamic environment
- The extended Soar for applying to the unknown environment
- The dynamic soar-rule generator based on reinforcement learning technique
- The argumentation-based negotiation among agents
- The argumentation method under uncertainty
- The knowledge-base for argumentation based on Bayesian-net/Markov Random Field(MRF)
- The rapid prototyping tool for constructing intelligent agent
- The rapper mechanism for logic programming language
- The novel script language for simple-use of logic programming language via our rapper mechanism
The solution for dealing with the solution stability in Dynamic Constraint Satisfaction Problem
Scheduling has been an important research field in Artificial
Intelligence. Because typical scheduling problems could be
modeled as Constraint Satisfaction Problems‾(CSP), several
constraint satisfaction techniques have been proposed. In
order to handle the different levels of importance of the
constraints, solving the problems via Valued Constraint
Satisfaction Problems‾(VCSP) is an promising approach.
However, there exists the case where an unexpected events
which might require a sudden change in the obtained schedule,
i.e., the case with dynamic changes in scheduling problems.
In this research, we describe such dynamic scheduling problem
as Dynamic Valued Constraint Satisfaction Problems‾(DyVCSP)
in which constraints would changes dynamically. Generally,
it is undesirable to determine vastly modified schedule even
if re-scheduling is needed. A new schedule should be close
to the current one as much as possible. In order to obtain
stable solutions, we propose the methodology to maintain
pieces of the current schedule via the use of temporal soft
constraints, which is explicitly penalizing the changes from
the current schedule. In this research, we focus on the
nurse scheduling for applying our method.
The dynamic soar-rule generator based on reinforcement learning technique
We present a Soar architecture agent using reinforcement
learning that has two features: 1) making a plan based on the
Soar architecture, and 2) resolving an impasse with
reinforcement learning. A Soar agent normally makes a plan
with production rules in recognition memory. When a Soar
agent cannot solve a problem with its recognition memory, it
generates an impasse. To resolve an impasse, a Soar agent
autonomously starts a reinforcement learning unit and
generates new production rules from the result of the
unit. The agent generates production rules with two steps: 1)
extracting association rules from a Q-function that is
learned with a reinforcement learning method, and 2) deriving
production rules from association rules with
EBL(Explanation-Based Learning) like method. The second step
uses a user ’s background knowledge to reduce the number of
production rules, and to facilitate comprehending what rules
has been derived. Generated production rules are stored in
the recognition memory. We have implemented a problem solver
for the container task environment to show how effectively
our Soar agent can work.
|
 |
|
publications |
- Hiromitsu Hattori, Toramatsu Shintani, Atsushi Isomura, Takayuki
Ito, and Tadachika Ozono, "Stable Solutions dealing with
Dynamics in Scheduling based on Dynamic Constraint Satisfaction
Problems ", The 8th Pacific Rim International Conference
on Artificial Intelligence (PRICAI-2004) (
in submission)
- Yoshinobu Bochi, Toramatsu Shintani, Takayuki Ito, and Tadachika
Ozono, "An Approach to Implementing Soar Agents Using Reinforcement
Learning", The 8th Pacific Rim International Conference on
Artificial Intelligence (PRICAI-2004) (
in submission)
|
Shintani Lab. 2004 |
|