Abstract
In this paper a condence measure is considered for an
agent who tries to keep a probabilistic model of her environment
of action. The measure is meant to capture only
one factor of the agent's doubt – namely, the issue whether
the agent has been able to collect a sufcient number of
observations. In this case stability of the agent's current
knowledge may give some clue about the trust she can put
in the model – indeed, some researchers from the eld of
probability theory suggest that such condence should be
based on the variance of the model (over time).
In this paper two different measures are proposed, both
based on aggregate variance of the estimator provided by
the learning process. The way the measures work is investigated
through some simple experiments with simulated
software agents. It turns out that an agent can benet from
using such measures as means for 'self-reection'. The
simulations suggest that the agent's condence should re-
ect the deviation of her knowledge from the reality. They
also show that it can be sometimes captured using very simple
methods: a measure proposed by Wang is tested in this
context, and it works seldom worse than the variance-based
measures, although it seems completely ad hoc and not well
suited for this particular setting of experiments at the rst
sight
Original language | Undefined |
---|---|
Pages | 39-44 |
Number of pages | 6 |
Publication status | Published - 2002 |
Event | 12th Belgian-Dutch Conference on Machine Learning, BeneLearn 2002 - University of Utrecht, Utrecht, Netherlands Duration: 4 Dec 2002 → 4 Dec 2002 Conference number: 12 |
Conference
Conference | 12th Belgian-Dutch Conference on Machine Learning, BeneLearn 2002 |
---|---|
Country/Territory | Netherlands |
City | Utrecht |
Period | 4/12/02 → 4/12/02 |
Keywords
- EWI-6656
- HMI-IA: Intelligent Agents
- IR-63359