AsistO: Un Sistema de Recomendaciones basado en MDPs Cualitativos para la Operación de Plantas Generadoras

2010 
THIS PAPER PROPOSES A NOVEL AND PRACTICAL MODEL-BASED LEARNING APPROACH WITH ITERATIVE REFINEMENT FOR SOLVING CONTINUOUS (AND HYBRID) MARKOV DECISION PROCESSES. INITIALLY, AN APPROXIMATE MODEL IS LEARNED USING CONVENTIONAL SAMPLING METHODS AND SOLVED TO OBTAIN A POLICY. ITERATIVELY, THE APPROXIMATE MODEL IS REFINED USING VARIANCE IN THE UTILITY VALUES AS PARTITION CRITERION. IN THE LEARNING PHASE, INITIAL REWARD AND TRANSITION FUNCTIONS ARE OBTAINED BY SAMPLING THE STATEACTION SPACE. THE SAMPLES ARE USED TO INDUCE A DECISION TREE PREDICTING REWARD VALUES FROM WHICH AN INITIAL PARTITION OF THE STATE SPACE IS BUILT. THE SAMPLES ARE ALSO USED TO INDUCE A FACTORED MDP. THE STATE ABSTRACTION IS THEN REFINED BY SPLITTING STATES ONLY WHERE THE SPLIT IS LOCALLY IMPORTANT. THE MAIN CONTRIBUTIONS OF THIS PAPER ARE THE USE OF SAMPLING TO CONSTRUCT AN ABSTRACTION, AND A LOCAL REFINEMENT PROCESS OF THE STATE ABSTRACTION BASED ON UTILITY VARIANCE. THE PROPOSED TECHNIQUE WAS TESTED IN ASISTO, AN INTELLIGENT RECOMMENDER SYSTEM FOR POWER PLANT OPERATION, WHERE WE SOLVED TWO VERSIONS OF A COMPLEX HYBRID CONTINUOUS-DISCRETE PROBLEM. WE SHOW HOW OUR TECHNIQUE APPROXIMATES A SOLUTION EVEN IN CASES WHERE STANDARD METHODS EXPLODE COMPUTATIONALLY.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    20
    References
    0
    Citations
    NaN
    KQI
    []