Datasize-Based Confidence Measure for a Learning Agent

M. Wiering (Editor), W.J. Jamroga

    Research output: Contribution to conferencePaperpeer-review


    In this paper a condence measure is considered for an agent who tries to keep a probabilistic model of her environment of action. The measure is meant to capture only one factor of the agent's doubt – namely, the issue whether the agent has been able to collect a sufcient number of observations. In this case stability of the agent's current knowledge may give some clue about the trust she can put in the model – indeed, some researchers from the eld of probability theory suggest that such condence should be based on the variance of the model (over time). In this paper two different measures are proposed, both based on aggregate variance of the estimator provided by the learning process. The way the measures work is investigated through some simple experiments with simulated software agents. It turns out that an agent can benet from using such measures as means for 'self-reection'. The simulations suggest that the agent's condence should re- ect the deviation of her knowledge from the reality. They also show that it can be sometimes captured using very simple methods: a measure proposed by Wang is tested in this context, and it works seldom worse than the variance-based measures, although it seems completely ad hoc and not well suited for this particular setting of experiments at the rst sight
    Original languageUndefined
    Number of pages6
    Publication statusPublished - 2002
    Event12th Belgian-Dutch Conference on Machine Learning, BeneLearn 2002 - University of Utrecht, Utrecht, Netherlands
    Duration: 4 Dec 20024 Dec 2002
    Conference number: 12


    Conference12th Belgian-Dutch Conference on Machine Learning, BeneLearn 2002


    • EWI-6656
    • HMI-IA: Intelligent Agents
    • IR-63359

    Cite this