2008 |
56 | | Sridhar Mahadevan:
Fast Spectral Learning using Lanczos Eigenspace Projections.
AAAI 2008: 1472-1475 |
55 | EE | Chang Wang,
Sridhar Mahadevan:
Manifold alignment using Procrustes analysis.
ICML 2008: 1120-1127 |
2007 |
54 | | Jeffrey Johns,
Sridhar Mahadevan,
Chang Wang:
Compact Spectral Bases for Value Function Approximation Using Kronecker Factorization.
AAAI 2007: 559-564 |
53 | | Ivon Arroyo,
Kimberly Ferguson,
Jeffrey Johns,
Toby Dragon,
Hasmik Meheranian,
Don Fisher,
Andrew G. Barto,
Sridhar Mahadevan,
Beverly Park Woolf:
Repairing Disengagement With Non-Invasive Interventions.
AIED 2007: 195-202 |
52 | | Sridhar Mahadevan,
Sarah Osentoski,
Jeffrey Johns,
Kimberly Ferguson,
Chang Wang:
Learning to Plan Using Harmonic Analysis of Diffusion Models.
ICAPS 2007: 224-231 |
51 | EE | Jeffrey Johns,
Sridhar Mahadevan:
Constructing basis functions from directed graphs for value function approximation.
ICML 2007: 385-392 |
50 | EE | Sridhar Mahadevan:
Adaptive mesh compression in 3D computer graphics using multiscale manifold learning.
ICML 2007: 585-592 |
49 | EE | Sarah Osentoski,
Sridhar Mahadevan:
Learning state-action basis functions for hierarchical MDPs.
ICML 2007: 705-712 |
2006 |
48 | | Sridhar Mahadevan,
Mauro Maggioni,
Kimberly Ferguson,
Sarah Osentoski:
Learning Representation and Control in Continuous Markov Decision Processes.
AAAI 2006 |
47 | EE | Mauro Maggioni,
Sridhar Mahadevan:
Fast direct policy evaluation using multiscale analysis of Markov diffusion processes.
ICML 2006: 601-608 |
46 | EE | Kimberly Ferguson,
Ivon Arroyo,
Sridhar Mahadevan,
Beverly Park Woolf,
Andrew G. Barto:
Improving Intelligent Tutoring Systems: Using Expectation Maximization to Learn Student Skill Levels.
Intelligent Tutoring Systems 2006: 453-462 |
45 | EE | Jeffrey Johns,
Sridhar Mahadevan,
Beverly Park Woolf:
Estimating Student Proficiency Using an Item Response Theory Model.
Intelligent Tutoring Systems 2006: 473-480 |
44 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan,
Rajbala Makar:
Hierarchical multi-agent reinforcement learning.
Autonomous Agents and Multi-Agent Systems 13(2): 197-229 (2006) |
2005 |
43 | | Sridhar Mahadevan:
Samuel Meets Amarel: Automating Value Function Approximation Using Global State Space Analysis.
AAAI 2005: 1000-1005 |
42 | | Jeffrey Johns,
Sridhar Mahadevan:
A Variational Learning Algorithm for the Abstract Hidden Markov Model.
AAAI 2005: 9-14 |
41 | EE | Sridhar Mahadevan:
Proto-value functions: developmental reinforcement learning.
ICML 2005: 553-560 |
40 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Coarticulation: an approach for generating concurrent plans in Markov decision processes.
ICML 2005: 720-727 |
39 | EE | Sridhar Mahadevan,
Mauro Maggioni:
Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions.
NIPS 2005 |
38 | EE | Sridhar Mahadevan:
Representation Policy Iteration.
UAI 2005: 372-379 |
2004 |
37 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Learning to Communicate and Act Using Hierarchical Reinforcement Learning.
AAMAS 2004: 1114-1121 |
36 | | Suchi Saria,
Sridhar Mahadevan:
Probabilistic Plan Recognition in Multiagent Systems.
ICAPS 2004: 287-296 |
35 | EE | Khashayar Rohanimanesh,
Robert Platt Jr.,
Sridhar Mahadevan,
Roderic A. Grupen:
Coarticulation in Markov Decision Processes.
NIPS 2004 |
2003 |
34 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Hierarchical Policy Gradient Algorithms.
ICML 2003: 226-233 |
33 | EE | Andrew G. Barto,
Sridhar Mahadevan:
Recent Advances in Hierarchical Reinforcement Learning.
Discrete Event Dynamic Systems 13(1-2): 41-77 (2003) |
32 | EE | Andrew G. Barto,
Sridhar Mahadevan:
Recent Advances in Hierarchical Reinforcement Learning.
Discrete Event Dynamic Systems 13(4): 341-379 (2003) |
2002 |
31 | EE | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
A multiagent reinforcement learning algorithm by dynamically merging markov decision processes.
AAMAS 2002: 845-846 |
30 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Hierarchically Optimal Average Reward Reinforcement Learning.
ICML 2002: 195-202 |
29 | | Georgios Theocharous,
Sridhar Mahadevan:
Approximate Planning with Hierarchical Partially Observable Markov Decision Process Models for Robot Navigation.
ICRA 2002: 1347-1352 |
28 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Learning to Take Concurrent Actions.
NIPS 2002: 1619-1626 |
27 | EE | Sridhar Mahadevan:
Spatiotemporal Abstraction of Stochastic Sequential Processes.
SARA 2002: 33-50 |
2001 |
26 | EE | Rajbala Makar,
Sridhar Mahadevan,
Mohammad Ghavamzadeh:
Hierarchical multi-agent reinforcement learning.
Agents 2001: 246-253 |
25 | EE | Silviu Minut,
Sridhar Mahadevan:
A reinforcement learning model of selective visual attention.
Agents 2001: 457-464 |
24 | | Mohammad Ghavamzadeh,
Sridhar Mahadevan:
Continuous-Time Hierarchical Reinforcement Learning.
ICML 2001: 186-193 |
23 | | Georgios Theocharous,
Khashayar Rohanimanesh,
Sridhar Mahadevan:
Learning Hierarchical Partially Observable Markov Decision Process Models for Robot Navigation.
ICRA 2001: 511-516 |
22 | EE | Khashayar Rohanimanesh,
Sridhar Mahadevan:
Decision-Theoretic Planning with Concurrent Temporally Extended Actions.
UAI 2001: 472-479 |
2000 |
21 | EE | Silviu Minut,
Sridhar Mahadevan,
John M. Henderson,
Fred C. Dyer:
Face Recognition Using Foveal Vision.
Biologically Motivated Computer Vision 2000: 424-433 |
20 | | Natalia Hernandez-Gardiol,
Sridhar Mahadevan:
Hierarchical Memory-Based Reinforcement Learning.
NIPS 2000: 1047-1053 |
1999 |
19 | | Gang Wang,
Sridhar Mahadevan:
Hierarchical Optimization of Policy-Coupled Semi-Markov Decision Processes.
ICML 1999: 464-473 |
1998 |
18 | | Sridhar Mahadevan,
Georgios Theocharous:
Optimizing Production Manufacturing Using Reinforcement Learning.
FLAIRS Conference 1998: 372-377 |
17 | | Sridhar Mahadevan,
Georgios Theocharous,
Nikfar Khaleeli:
Rapid Concept Learning for Mobile Robots.
Auton. Robots 5(3-4): 239-251 (1998) |
16 | | Sridhar Mahadevan,
Georgios Theocharous,
Nikfar Khaleeli:
Rapid Concept Learning for Mobile Robots.
Machine Learning 31(1-3): 7-27 (1998) |
1996 |
15 | | Sridhar Mahadevan:
An Average-Reward Reinforcement Learning Algorithm for Computing Bias-Optimal Policies.
AAAI/IAAI, Vol. 1 1996: 875-880 |
14 | | Sridhar Mahadevan:
Sensitive Discount Optimality: Unifying Discounted and Average Reward Reinforcement Learning.
ICML 1996: 328-336 |
13 | | Sridhar Mahadevan,
Leslie Pack Kaelbling:
The National Science Foundation Workshop on Reinforcement Learning.
AI Magazine 17(4): 89-93 (1996) |
12 | | Sridhar Mahadevan:
Average Reward Reinforcement Learning: Foundations, Algorithms, and Empirical Results.
Machine Learning 22(1-3): 159-195 (1996) |
1994 |
11 | | Sridhar Mahadevan:
To Discount or Not to Discount in Reinforcement Learning: A Case Study Comparing R Learning and Q Learning.
ICML 1994: 164-172 |
10 | | Sridhar Mahadevan,
Prasad Tadepalli:
Quantifying Prior Determination Knowledge Using the PAC Learning Model.
Machine Learning 17(1): 69-105 (1994) |
1993 |
9 | | Sridhar Mahadevan,
Tom M. Mitchell,
Jack Mostow,
Louis I. Steinberg,
Prasad Tadepalli:
An Apprentice-Based Approach to Knowledge Acquisition.
Artif. Intell. 64(1): 1-52 (1993) |
1992 |
8 | | Sridhar Mahadevan:
Enhancing Transfer in Reinforcement Learning by Building Stochastic Models of Robot Actions.
ML 1992: 290-299 |
7 | | Sridhar Mahadevan,
Jonathan Connell:
Automatic Programming of Behavior-Based Robots Using Reinforcement Learning.
Artif. Intell. 55(2): 311-365 (1992) |
1991 |
6 | | Sridhar Mahadevan,
Jonathan Connell:
Automatic Programming of Behavior-Based Robots Using Reinforcement Learning.
AAAI 1991: 768-773 |
5 | | Sridhar Mahadevan,
Jonathan Connell:
Scaling Reinforcement Learning to Robotics by Exploiting the Subsumption Architecture.
ML 1991: 328-332 |
1989 |
4 | | Sridhar Mahadevan:
Using Determinations in EBL: A Solution to the incomplete Theory Problem.
ML 1989: 320-325 |
1988 |
3 | | Sridhar Mahadevan,
Prasad Tadepalli:
On the Tractability of Learning from Incomplete Theories.
ML 1988: 235-241 |
1985 |
2 | | Tom M. Mitchell,
Sridhar Mahadevan,
Louis I. Steinberg:
LEAP: A Learning Apprentice for VLSl Design.
IJCAI 1985: 573-580 |
1 | | Sridhar Mahadevan:
Verification-based Learning: A Generalized Strategy for Inferring Problem-Reduction Methods.
IJCAI 1985: 616-623 |