| 2008 |
| 56 | | Sridhar Mahadevan:
Fast Spectral Learning using Lanczos Eigenspace Projections.
AAAI 2008: 1472-1475 |
| 55 | EE | Chang Wang,
Sridhar Mahadevan:
Manifold alignment using Procrustes analysis.
ICML 2008: 1120-1127 |
| 2007 |
| 54 | | Jeffrey Johns,
Sridhar Mahadevan,
Chang Wang:
Compact Spectral Bases for Value Function Approximation Using Kronecker Factorization.
AAAI 2007: 559-564 |
| 53 | | Ivon Arroyo,
Kimberly Ferguson,
Jeffrey Johns,
Toby Dragon,
Hasmik Meheranian,
Don Fisher,
Andrew G. Barto,
Sridhar Mahadevan,
Beverly Park Woolf:
Repairing Disengagement With Non-Invasive Interventions.
AIED 2007: 195-202 |
| 52 | | Sridhar Mahadevan,
Sarah Osentoski,
Jeffrey Johns,
Kimberly Ferguson,
Chang Wang:
Learning to Plan Using Harmonic Analysis of Diffusion Models.
ICAPS 2007: 224-231 |
| 51 | EE | Jeffrey Johns,
Sridhar Mahadevan:
Constructing basis functions from directed graphs for value function approximation.
ICML 2007: 385-392 |
| 50 | EE | Sridhar Mahadevan:
Adaptive mesh compression in 3D computer graphics using multiscale manifold learning.
ICML 2007: 585-592 |
| 49 | EE | Sarah Osentoski,
Sridhar Mahadevan:
Learning state-action basis functions for hierarchical MDPs.
ICML 2007: 705-712 |
| 2006 |
| 48 | | Sridhar Mahadevan,
Mauro Maggioni,
Kimberly Ferguson,
Sarah Osentoski:
Learning Representation and Control in Continuous Markov Decision Processes.
AAAI 2006 |
| 47 | EE | Mauro Maggioni,
Sridhar Mahadevan:
Fast direct policy evaluation using multiscale analysis of Markov diffusion processes.
ICML 2006: 601-608 |
| 46 | EE | Kimberly Ferguson,
Ivon Arroyo,
Sridhar Mahadevan,
Beverly Park Woolf,
Andrew G. Barto:
Improving Intelligent Tutoring Systems: Using Expectation Maximization to Learn Student Skill Levels.
Intelligent Tutoring Systems 2006: 453-462 |
| 45 | EE | Jeffrey Johns,
Sridhar Mahadevan,
Beverly Park Woolf:
Estimating Student Proficiency Using an Item Response Theory Model.
Intelligent Tutoring Systems 2006: 473-480 |
| 44 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan,
Rajbala Makar:
Hierarchical multi-agent reinforcement learning.
Autonomous Agents and Multi-Agent Systems 13(2): 197-229 (2006) |
| 2005 |
| 43 | | Sridhar Mahadevan:
Samuel Meets Amarel: Automating Value Function Approximation Using Global State Space Analysis.
AAAI 2005: 1000-1005 |
| 42 | | Jeffrey Johns,
Sridhar Mahadevan:
A Variational Learning Algorithm for the Abstract Hidden Markov Model.
AAAI 2005: 9-14 |
| 41 | EE | Sridhar Mahadevan:
Proto-value functions: developmental reinforcement learning.
ICML 2005: 553-560 |
| 40 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Coarticulation: an approach for generating concurrent plans in Markov decision processes.
ICML 2005: 720-727 |
| 39 | EE | Sridhar Mahadevan,
Mauro Maggioni:
Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions.
NIPS 2005 |
| 38 | EE | Sridhar Mahadevan:
Representation Policy Iteration.
UAI 2005: 372-379 |
| 2004 |
| 37 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Learning to Communicate and Act Using Hierarchical Reinforcement Learning.
AAMAS 2004: 1114-1121 |
| 36 | | Suchi Saria,
Sridhar Mahadevan:
Probabilistic Plan Recognition in Multiagent Systems.
ICAPS 2004: 287-296 |
| 35 | EE | Khashayar Rohanimanesh,
Robert Platt Jr.,
Sridhar Mahadevan,
Roderic A. Grupen:
Coarticulation in Markov Decision Processes.
NIPS 2004 |
| 2003 |
| 34 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Hierarchical Policy Gradient Algorithms.
ICML 2003: 226-233 |
| 33 | EE | Andrew G. Barto,
Sridhar Mahadevan:
Recent Advances in Hierarchical Reinforcement Learning.
Discrete Event Dynamic Systems 13(1-2): 41-77 (2003) |
| 32 | EE | Andrew G. Barto,
Sridhar Mahadevan:
Recent Advances in Hierarchical Reinforcement Learning.
Discrete Event Dynamic Systems 13(4): 341-379 (2003) |
| 2002 |
| 31 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
A multiagent reinforcement learning algorithm by dynamically merging markov decision processes.
AAMAS 2002: 845-846 |
| 30 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Hierarchically Optimal Average Reward Reinforcement Learning.
ICML 2002: 195-202 |
| 29 | | Georgios Theocharous,
Sridhar Mahadevan:
Approximate Planning with Hierarchical Partially Observable Markov Decision Process Models for Robot Navigation.
ICRA 2002: 1347-1352 |
| 28 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Learning to Take Concurrent Actions.
NIPS 2002: 1619-1626 |
| 27 | EE | Sridhar Mahadevan:
Spatiotemporal Abstraction of Stochastic Sequential Processes.
SARA 2002: 33-50 |
| 2001 |
| 26 | EE | Rajbala Makar,
Sridhar Mahadevan,
Mohammad Ghavamzadeh:
Hierarchical multi-agent reinforcement learning.
Agents 2001: 246-253 |
| 25 | EE | Silviu Minut,
Sridhar Mahadevan:
A reinforcement learning model of selective visual attention.
Agents 2001: 457-464 |
| 24 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Continuous-Time Hierarchical Reinforcement Learning.
ICML 2001: 186-193 |
| 23 | | Georgios Theocharous,
Khashayar Rohanimanesh,
Sridhar Mahadevan:
Learning Hierarchical Partially Observable Markov Decision Process Models for Robot Navigation.
ICRA 2001: 511-516 |
| 22 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Decision-Theoretic Planning with Concurrent Temporally Extended Actions.
UAI 2001: 472-479 |
| 2000 |
| 21 | EE | Silviu Minut,
Sridhar Mahadevan,
John M. Henderson,
Fred C. Dyer:
Face Recognition Using Foveal Vision.
Biologically Motivated Computer Vision 2000: 424-433 |
| 20 | | Natalia Hernandez-Gardiol,
Sridhar Mahadevan:
Hierarchical Memory-Based Reinforcement Learning.
NIPS 2000: 1047-1053 |
| 1999 |
| 19 | | Gang Wang,
Sridhar Mahadevan:
Hierarchical Optimization of Policy-Coupled Semi-Markov Decision Processes.
ICML 1999: 464-473 |
| 1998 |
| 18 | | Sridhar Mahadevan,
Georgios Theocharous:
Optimizing Production Manufacturing Using Reinforcement Learning.
FLAIRS Conference 1998: 372-377 |
| 17 | | Sridhar Mahadevan,
Georgios Theocharous,
Nikfar Khaleeli:
Rapid Concept Learning for Mobile Robots.
Auton. Robots 5(3-4): 239-251 (1998) |
| 16 | | Sridhar Mahadevan,
Georgios Theocharous,
Nikfar Khaleeli:
Rapid Concept Learning for Mobile Robots.
Machine Learning 31(1-3): 7-27 (1998) |
| 1996 |
| 15 | | Sridhar Mahadevan:
An Average-Reward Reinforcement Learning Algorithm for Computing Bias-Optimal Policies.
AAAI/IAAI, Vol. 1 1996: 875-880 |
| 14 | | Sridhar Mahadevan:
Sensitive Discount Optimality: Unifying Discounted and Average Reward Reinforcement Learning.
ICML 1996: 328-336 |
| 13 | | Sridhar Mahadevan,
Leslie Pack Kaelbling:
The National Science Foundation Workshop on Reinforcement Learning.
AI Magazine 17(4): 89-93 (1996) |
| 12 | | Sridhar Mahadevan:
Average Reward Reinforcement Learning: Foundations, Algorithms, and Empirical Results.
Machine Learning 22(1-3): 159-195 (1996) |
| 1994 |
| 11 | | Sridhar Mahadevan:
To Discount or Not to Discount in Reinforcement Learning: A Case Study Comparing R Learning and Q Learning.
ICML 1994: 164-172 |
| 10 | | Sridhar Mahadevan,
Prasad Tadepalli:
Quantifying Prior Determination Knowledge Using the PAC Learning Model.
Machine Learning 17(1): 69-105 (1994) |
| 1993 |
| 9 | | Sridhar Mahadevan,
Tom M. Mitchell,
Jack Mostow,
Louis I. Steinberg,
Prasad Tadepalli:
An Apprentice-Based Approach to Knowledge Acquisition.
Artif. Intell. 64(1): 1-52 (1993) |
| 1992 |
| 8 | | Sridhar Mahadevan:
Enhancing Transfer in Reinforcement Learning by Building Stochastic Models of Robot Actions.
ML 1992: 290-299 |
| 7 | | Sridhar Mahadevan,
Jonathan Connell:
Automatic Programming of Behavior-Based Robots Using Reinforcement Learning.
Artif. Intell. 55(2): 311-365 (1992) |
| 1991 |
| 6 | | Sridhar Mahadevan,
Jonathan Connell:
Automatic Programming of Behavior-Based Robots Using Reinforcement Learning.
AAAI 1991: 768-773 |
| 5 | | Sridhar Mahadevan,
Jonathan Connell:
Scaling Reinforcement Learning to Robotics by Exploiting the Subsumption Architecture.
ML 1991: 328-332 |
| 1989 |
| 4 | | Sridhar Mahadevan:
Using Determinations in EBL: A Solution to the incomplete Theory Problem.
ML 1989: 320-325 |
| 1988 |
| 3 | | Sridhar Mahadevan,
Prasad Tadepalli:
On the Tractability of Learning from Incomplete Theories.
ML 1988: 235-241 |
| 1985 |
| 2 | | Tom M. Mitchell,
Sridhar Mahadevan,
Louis I. Steinberg:
LEAP: A Learning Apprentice for VLSl Design.
IJCAI 1985: 573-580 |
| 1 | | Sridhar Mahadevan:
Verification-based Learning: A Generalized Strategy for Inferring Problem-Reduction Methods.
IJCAI 1985: 616-623 |