Pieter Abbeel
Machine learning researcher at UC-Berkeley
Why Is Pieter Abbeel Influential?
(Suggest an Edit or Addition)According to Wikipedia, Pieter Abbeel is a professor of electrical engineering and computer sciences, Director of the Berkeley Robot Learning Lab, and co-director of the Berkeley AI Research Lab at the University of California, Berkeley. He is also the co-founder of covariant.ai, a venture-funded start-up that aims to teach robots new, complex skills, and co-founder of Gradescope, an online grading system that has been implemented in over 500 universities nationwide. He is best known for his cutting-edge research in robotics and machine learning, particularly in deep reinforcement learning. In 2021, he joined AIX Ventures as an Investment Partner. AIX Ventures is a venture capital fund that invests in artificial intelligence startups.
Pieter Abbeel's Published Works
Published Works
- Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks (2017) (7450)
- Trust Region Policy Optimization (2015) (4825)
- Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor (2018) (4170)
- InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets (2016) (3593)
- Apprenticeship learning via inverse reinforcement learning (2004) (2888)
- End-to-End Training of Deep Visuomotor Policies (2015) (2811)
- Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments (2017) (2478)
- High-Dimensional Continuous Control Using Generalized Advantage Estimation (2015) (2133)
- Domain randomization for transferring deep neural networks from simulation to the real world (2017) (2025)
- Denoising Diffusion Probabilistic Models (2020) (1938)
- Hindsight Experience Replay (2017) (1487)
- Benchmarking Deep Reinforcement Learning for Continuous Control (2016) (1421)
- Soft Actor-Critic Algorithms and Applications (2018) (1169)
- A Simple Neural Attentive Meta-Learner (2017) (1032)
- Reinforcement Learning with Deep Energy-Based Policies (2017) (905)
- Sim-to-Real Transfer of Robotic Control with Dynamics Randomization (2017) (896)
- Discriminative Probabilistic Models for Relational Data (2002) (809)
- RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning (2016) (761)
- Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization (2016) (750)
- Constrained Policy Optimization (2017) (747)
- A Survey of Research on Cloud Robotics and Automation (2015) (701)
- An Application of Reinforcement Learning to Aerobatic Helicopter Flight (2006) (660)
- CURL: Contrastive Unsupervised Representations for Reinforcement Learning (2020) (643)
- VIME: Variational Information Maximizing Exploration (2016) (619)
- Autonomous Helicopter Aerobatics through Apprenticeship Learning (2010) (608)
- DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills (2018) (606)
- Adversarial Attacks on Neural Network Policies (2017) (598)
- Adaptive Control Processes (2010) (586)
- Variational Lossy Autoencoder (2016) (576)
- Motion planning with sequential convex optimization and convex collision checking (2014) (566)
- #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning (2016) (562)
- Overcoming Exploration in Reinforcement Learning with Demonstrations (2017) (560)
- Value Iteration Networks (2016) (560)
- Link Prediction in Relational Data (2003) (540)
- One-Shot Imitation Learning (2017) (522)
- Emergence of Grounded Compositional Language in Multi-Agent Populations (2017) (521)
- An Algorithmic Perspective on Imitation Learning (2018) (514)
- The YCB object and Model set: Towards common benchmarks for manipulation research (2015) (511)
- Learning to Poke by Poking: Experiential Learning of Intuitive Physics (2016) (502)
- Parameter Space Noise for Exploration (2017) (488)
- Bottleneck Transformers for Visual Recognition (2021) (481)
- Deep spatial autoencoders for visuomotor learning (2015) (470)
- Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics (2014) (459)
- Cooperative Inverse Reinforcement Learning (2016) (455)
- Decision Transformer: Reinforcement Learning via Sequence Modeling (2021) (433)
- Combined task and motion planning through an extensible planner-independent interface layer (2014) (429)
- Finding Locally Optimal, Collision-Free Trajectories with Sequential Convex Optimization (2013) (428)
- Evaluating Protein Transfer Learning with TAPE (2019) (425)
- Cloth grasp point detection based on multiple-view geometric cues with application to robotic towel folding (2010) (423)
- Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models (2015) (422)
- One-Shot Visual Imitation Learning via Meta-Learning (2017) (416)
- Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation (2017) (408)
- The limits and potentials of deep learning for robotics (2018) (406)
- Efficient L1 Regularized Logistic Regression (2006) (403)
- Learning with Opponent-Learning Awareness (2017) (385)
- Learning deep control policies for autonomous aerial vehicles with MPC-guided policy search (2015) (383)
- Automatic Goal Generation for Reinforcement Learning Agents (2017) (369)
- Reinforcement Learning with Augmented Data (2020) (368)
- Stochastic Adversarial Video Prediction (2018) (368)
- Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning (2018) (367)
- Gradient Estimation Using Stochastic Computation Graphs (2015) (336)
- Model-Ensemble Trust-Region Policy Optimization (2018) (334)
- Reverse Curriculum Generation for Reinforcement Learning (2017) (326)
- Benchmarking in Manipulation Research: Using the Yale-CMU-Berkeley Object and Model Set (2015) (323)
- Flow++: Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design (2019) (320)
- Learning contact-rich manipulation skills with guided policy search (2015) (313)
- Stochastic Neural Networks for Hierarchical Reinforcement Learning (2016) (308)
- Learning modular neural network policies for multi-task and multi-robot transfer (2016) (307)
- LQG-MP: Optimized path planning for robots with motion uncertainty and imperfect state information (2010) (306)
- BigBIRD: A large-scale 3D database of object instances (2014) (298)
- Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments (2017) (290)
- Meta Learning Shared Hierarchies (2017) (282)
- A Connection between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy-Based Models (2016) (279)
- One-Shot Imitation from Observing Humans via Domain-Adaptive Meta-Learning (2018) (277)
- Population Based Augmentation: Efficient Learning of Augmentation Policy Schedules (2019) (277)
- Imitation from Observation: Learning to Imitate Behaviors from Raw Video via Context Translation (2017) (269)
- Learning for control from multiple demonstrations (2008) (266)
- Equivalence Between Policy Gradients and Soft Q-Learning (2017) (263)
- Exploration and apprenticeship learning in reinforcement learning (2005) (262)
- Inverse Reward Design (2017) (261)
- Safe Exploration in Markov Decision Processes (2012) (260)
- Uncertainty-Aware Reinforcement Learning for Collision Avoidance (2017) (260)
- Meta-Reinforcement Learning of Structured Exploration Strategies (2018) (253)
- Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model (2019) (252)
- Asymmetric Actor Critic for Image-Based Robot Learning (2017) (250)
- A geometric approach to robotic laundry folding (2012) (246)
- Learning the Dynamics of Arterial Traffic From Probe Data Using a Dynamic Bayesian Network (2012) (244)
- Using inaccurate models in reinforcement learning (2006) (243)
- Benchmarking Model-Based Reinforcement Learning (2019) (236)
- Superhuman performance of surgical tasks by robots using iterative learning from human-guided demonstrations (2010) (235)
- Combining self-supervised learning and imitation for vision-based rope manipulation (2017) (230)
- Yale-CMU-Berkeley dataset for robotic manipulation research (2017) (229)
- Self-Supervised Deep Reinforcement Learning with Generalized Computation Graphs for Robot Navigation (2017) (228)
- Estimating arterial traffic conditions using sparse probe data (2010) (217)
- LQG-MP: Optimized path planning for robots with motion uncertainty and imperfect state information (2011) (214)
- Learning Invariant Feature Spaces to Transfer Skills with Reinforcement Learning (2017) (212)
- Benchmarking in Manipulation Research: The YCB Object and Model Set and Benchmarking Protocols (2015) (209)
- Planning to Explore via Self-Supervised World Models (2020) (207)
- Transfer from Simulation to Real World through Learning Deep Inverse Dynamics Model (2016) (204)
- Decoupling Representation Learning from Reinforcement Learning (2020) (198)
- A robot path planning framework that learns from experience (2012) (194)
- Third-Person Imitation Learning (2017) (191)
- Evolved Policy Gradients (2018) (184)
- Composable Deep Reinforcement Learning for Robotic Manipulation (2018) (184)
- PixelSNAIL: An Improved Autoregressive Generative Model (2017) (182)
- The Path Inference Filter: Model-Based Low-Latency Map Matching of Probe Vehicle Data (2011) (181)
- Meta-Learning with Temporal Convolutions (2017) (179)
- A Berkeley View of Systems Challenges for AI (2017) (171)
- SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning (2018) (171)
- SFV: Reinforcement Learning of Physical Skills from Videos (2018) (171)
- Model-Based Reinforcement Learning via Meta-Policy Optimization (2018) (165)
- Learning Factor Graphs in Polynomial Time and Sample Complexity (2006) (165)
- Path and travel time inference from GPS probe vehicle data (2009) (163)
- Information-Theoretic Planning with Trajectory Optimization for Dense 3D Mapping (2015) (163)
- Backprop KF: Learning Discriminative Deterministic State Estimators (2016) (162)
- Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow (2018) (161)
- Bringing clothing into desired configurations with limited perception (2011) (161)
- Tracking deformable objects with point clouds (2013) (158)
- ProMP: Proximal Meta-Policy Search (2018) (157)
- Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion (2007) (157)
- Learning by observation for surgical subtasks: Multilateral cutting of 3D viscoelastic and 2D Orthotropic Tissue Phantoms (2015) (156)
- Latent Space Policies for Hierarchical Reinforcement Learning (2018) (151)
- On the Utility of Learning about Humans for Human-AI Coordination (2019) (150)
- Deep learning helicopter dynamics models (2015) (144)
- One-shot learning of manipulation skills with online dynamics adaptation and neural network priors (2015) (142)
- Learning Plannable Representations with Causal InfoGAN (2018) (142)
- Putting NeRF on a Diet: Semantically Consistent Few-Shot View Synthesis (2021) (140)
- Pretrained Transformers as Universal Computation Engines (2021) (138)
- Discriminative Training of Kalman Filters (2005) (138)
- Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents (2022) (138)
- A textured object recognition pipeline for color and depth image data (2012) (137)
- Apprenticeship learning for motion planning with application to parking lot navigation (2008) (135)
- Domain Randomization and Generative Models for Robotic Grasping (2017) (135)
- Universal Planning Networks (2018) (135)
- MSA Transformer (2021) (135)
- Max-margin Classification of Data with Absent Features (2008) (134)
- Learning from Demonstrations Through the Use of Non-rigid Registration (2013) (134)
- MCP: Learning Composable Hierarchical Control with Multiplicative Compositional Policies (2019) (134)
- Grounding spatial relations for human-robot interaction (2013) (131)
- VideoGPT: Video Generation using VQ-VAE and Transformers (2021) (131)
- Learning dexterous manipulation for a soft robotic hand from human demonstrations (2016) (131)
- PLATO: Policy learning using adaptive trajectory optimization (2016) (127)
- Deep Unsupervised Cardinality Estimation (2019) (126)
- Transition state clustering: Unsupervised surgical trajectory segmentation for robot learning (2017) (124)
- Variance Reduction for Policy Gradient with Action-Dependent Factorized Baselines (2018) (122)
- Learning to Manipulate Deformable Objects without Demonstrations (2019) (121)
- Adapting Deep Visuomotor Representations with Weak Pairwise Constraints (2015) (121)
- Learning Predictive Representations for Deformable Objects Using Contrastive Estimation (2020) (121)
- Using Mobile Phones to Forecast Arterial Traffic through Statistical Learning (2010) (121)
- Learning force-based manipulation of deformable objects from multiple demonstrations (2015) (121)
- Zero-Shot Text-Guided Object Generation with Dream Fields (2021) (120)
- Variational Option Discovery Algorithms (2018) (118)
- Responsive Safety in Reinforcement Learning by PID Lagrangian Methods (2020) (115)
- Learning Robotic Assembly from CAD (2018) (115)
- Enabling robots to communicate their objectives (2017) (114)
- Goal-conditioned Imitation Learning (2019) (113)
- Parametrized shape models for clothing (2011) (110)
- Autonomous multilateral debridement with the Raven surgical robot (2014) (110)
- Self-Consistent Trajectory Autoencoder: Hierarchical Reinforcement Learning with Trajectory Embeddings (2018) (109)
- SUNRISE: A Simple Unified Framework for Ensemble Learning in Deep Reinforcement Learning (2020) (108)
- Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization (2019) (107)
- AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control (2021) (106)
- Accelerated Methods for Deep Reinforcement Learning (2018) (104)
- Construction of Cubic Structures with Quadrotor Teams (2012) (99)
- Apprenticeship learning for helicopter control (2009) (98)
- Reinforcement Learning on Variable Impedance Controller for High-Precision Robotic Assembly (2019) (98)
- BADGR: An Autonomous Self-Supervised Learning-Based Navigation System (2020) (97)
- Some Considerations on Learning to Explore via Meta-Reinforcement Learning (2018) (90)
- Generalization through Simulation: Integrating Simulated and Real Data into Deep Reinforcement Learning for Vision-Based Autonomous Flight (2019) (89)
- UCB EXPLORATION VIA Q-ENSEMBLES (2018) (87)
- Learning vehicular dynamics, with application to modeling helicopters (2005) (87)
- The Off-Switch Game (2016) (87)
- Learning Robotic Manipulation through Visual Planning and Acting (2019) (86)
- Scaling up Gaussian Belief Space Planning Through Covariance-Free Trajectory Optimization and Automatic Differentiation (2014) (86)
- Gravity-Based Robotic Cloth Folding (2010) (86)
- Learning Generalized Reactive Policies using Deep Neural Networks (2017) (86)
- Self-Supervised Policy Adaptation during Deployment (2020) (85)
- Deep reinforcement learning for tensegrity robot locomotion (2016) (85)
- Behavior From the Void: Unsupervised Active Pre-Training (2021) (85)
- rlpyt: A Research Code Base for Deep Reinforcement Learning in PyTorch (2019) (83)
- A case study of trajectory transfer through non-rigid registration for a simplified suturing scenario (2013) (83)
- Universal Planning Networks: Learning Generalizable Representations for Visuomotor Control (2018) (83)
- Towards Characterizing Divergence in Deep Q-Learning (2019) (80)
- Learning deep neural network policies with continuous memory states (2015) (79)
- Apprenticeship learning and reinforcement learning with application to robotic control (2008) (79)
- Scaling the mobile millennium system in the cloud (2011) (77)
- Reinforcement Learning for Robust Parameterized Locomotion Control of Bipedal Robots (2021) (77)
- AVID: Learning Multi-Stage Tasks via Pixel-Level Translation of Human Videos (2019) (76)
- Learning to Adapt: Meta-Learning for Model-Based Control (2018) (76)
- Deep Object-Centric Representations for Generalizable Robot Learning (2017) (72)
- Autonomous Autorotation of an RC Helicopter (2008) (71)
- On a Connection between Importance Sampling and the Likelihood Ratio Policy Gradient (2010) (71)
- Bit-Swap: Recursive Bits-Back Coding for Lossless Compression with Hierarchical Latent Variables (2019) (70)
- Contrastive Code Representation Learning (2020) (69)
- Learning Visual Servoing with Deep Features and Fitted Q-Iteration (2017) (69)
- Learning Visual Feature Spaces for Robotic Manipulation with Deep Spatial Autoencoders (2015) (68)
- Learning accurate kinematic control of cable-driven surgical robots using data cleaning and Gaussian Process Regression (2014) (67)
- Performance analysis and terrain classification for a legged robot over rough terrain (2012) (66)
- Curiosity-driven Exploration in Deep Reinforcement Learning via Bayesian Neural Networks (2016) (66)
- Establishing Appropriate Trust via Critical States (2018) (65)
- TSC-DL: Unsupervised trajectory segmentation of multi-modal surgical demonstrations with Deep Learning (2016) (63)
- Geometric Programming for Aircraft Design Optimization (2012) (63)
- A single-use haptic palpation probe for locating subcutaneous blood vessels in robot-assisted minimally invasive surgery (2015) (63)
- GP-GPIS-OPT: Grasp planning with shape uncertainty using Gaussian process implicit surfaces and Sequential Convex Programming (2015) (62)
- Physics-based trajectory optimization for grasping in cluttered environments (2015) (61)
- LQG-Based Planning, Sensing, and Control of Steerable Needles (2010) (60)
- A Framework for Efficient Robotic Manipulation (2020) (60)
- PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via Relabeling Experience and Unsupervised Pre-training (2021) (59)
- Gradescope: A Fast, Flexible, and Fair System for Scalable Assessment of Handwritten Work (2017) (59)
- Generalizing Skills with Semi-Supervised Reinforcement Learning (2016) (58)
- Automatic Curriculum Learning through Value Disagreement (2020) (58)
- Mastering Atari Games with Limited Data (2021) (57)
- Model-Augmented Actor-Critic: Backpropagating through Paths (2020) (57)
- One-Shot Hierarchical Imitation Learning of Compound Visuomotor Tasks (2018) (57)
- Multimodal blending for high-accuracy instance recognition (2013) (56)
- APS: Active Pretraining with Successor Features (2021) (56)
- Prediction and Control with Temporal Segment Models (2017) (56)
- Toward asymptotically optimal motion planning for kinodynamic systems using a two-point boundary value problem solver (2015) (56)
- State Entropy Maximization with Random Encoders for Efficient Exploration (2021) (55)
- EG-RRT: Environment-guided random trees for kinodynamic motion planning with uncertainty and obstacles (2011) (54)
- Guided search for task and motion plans using learned heuristics (2016) (54)
- Learning from the hindsight plan — Episodic MPC improvement (2016) (52)
- URLB: Unsupervised Reinforcement Learning Benchmark (2021) (50)
- Visual Imitation Made Easy (2020) (50)
- Combining model-based policy search with online model learning for control of physical humanoids (2016) (49)
- Active exploration using trajectory optimization for robotic grasping in the presence of occlusions (2015) (48)
- Robust Reinforcement Learning using Adversarial Populations (2020) (48)
- Mutual Alignment Transfer Learning (2017) (47)
- Guiding Policies with Language via Meta-Learning (2018) (47)
- Variational Information Maximizing Exploration (2016) (46)
- Deciphering the Role of a Coleopteran Steering Muscle via Free Flight Stimulation (2015) (46)
- Modeling and perception of deformable one-dimensional objects (2011) (45)
- Offline-to-Online Reinforcement Learning via Balanced Replay and Pessimistic Q-Ensemble (2021) (45)
- Guided Meta-Policy Search (2019) (44)
- Sub-policy Adaptation for Hierarchical Reinforcement Learning (2019) (43)
- Model-based reinforcement learning with parametrized physical models and optimism-driven exploration (2015) (42)
- Area Under Curve (2020) (42)
- Preferences Implicit in the State of the World (2018) (42)
- Perception for the manipulation of socks (2011) (41)
- Modular task and motion planning in belief space (2015) (41)
- Modular Architecture for StarCraft II with Deep Reinforcement Learning (2018) (41)
- Learning compound multi-step controllers under unknown dynamics (2015) (41)
- Risk Aversion in Markov Decision Processes via Near Optimal Chernoff Bounds (2012) (41)
- Generalized Hindsight for Reinforcement Learning (2020) (41)
- Learning first-order Markov models for control (2004) (41)
- DoorGym: A Scalable Door Opening Environment And Baseline Agent (2019) (41)
- DeepMimic (2018) (41)
- On the Feasibility of Learning, Rather than Assuming, Human Biases for Reward Inference (2019) (40)
- Parameterized maneuver learning for autonomous helicopter flight (2010) (40)
- Using Classical Planners for Tasks with Continuous Operators in Robotics (2013) (40)
- A Biological Micro Actuator: Graded and Closed-Loop Control of Insect Leg Motion by Electrical Stimulation of Muscles (2014) (40)
- Safer Classification by Synthesis (2017) (39)
- Sigma hulls for Gaussian belief space planning for imprecise articulated robots amid obstacles (2013) (39)
- Leveraging appearance priors in non-rigid registration, with application to manipulation of deformable objects (2015) (38)
- Real-World Robot Learning with Masked Visual Pre-training (2022) (37)
- Compression with Flows via Local Bits-Back Coding (2019) (37)
- BagNet: Berkeley Analog Generator with Layout Optimizer Boosted with Deep Neural Networks (2019) (36)
- Large-Scale Estimation in Cyberphysical Systems Using Streaming Data: A Case Study With Arterial Traffic Estimation (2013) (36)
- Auto-Tuned Sim-to-Real Transfer (2021) (36)
- SOLAR: Deep Structured Latent Representations for Model-Based Reinforcement Learning (2018) (36)
- Multi-armed bandit models for 2D grasp planning with uncertainty (2015) (35)
- Reset-free guided policy search: Efficient deep reinforcement learning with stochastic initial states (2016) (35)
- Optimism-driven exploration for nonlinear systems (2015) (34)
- Quasi-Direct Drive for Low-Cost Compliant Robotic Manipulation (2019) (34)
- Multimodal Masked Autoencoders Learn Transferable Representations (2022) (34)
- Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning (2022) (33)
- Global Motion Planning under Uncertain Motion, Sensing, and Environment Map (2012) (33)
- Learning from multiple demonstrations using trajectory-aware non-rigid registration with applications to deformable object manipulation (2015) (33)
- Hallucinative Topological Memory for Zero-Shot Visual Planning (2020) (33)
- Hybrid Discriminative-Generative Training via Contrastive Learning (2020) (33)
- Domain Randomization for Active Pose Estimation (2019) (33)
- DayDreamer: World Models for Physical Robot Learning (2022) (32)
- Sparse Graphical Memory for Robust Planning (2020) (32)
- Benchmarking in Manipulation Research (2015) (30)
- Reinforcement Learning with Action-Free Pre-Training from Videos (2022) (30)
- Gaussian belief space planning with discontinuities in sensing domains (2014) (30)
- Label and Link Prediction in Relational Data (2003) (29)
- Unmanned Aircraft Collision Avoidance Using Continuous-State POMDPs (2012) (28)
- A Linear Approximation for Graph-Based Simultaneous Localization and Mapping (2012) (28)
- Predicting initialization effectiveness for trajectory optimization (2014) (28)
- Data fitting with geometric-programming-compatible softmax functions (2016) (28)
- Generalization in Dexterous Manipulation via Geometry-Aware Multi-Task Learning (2021) (28)
- Motion planning and control of robotic manipulators on seaborne platforms (2011) (28)
- Unifying scene registration and trajectory optimization for learning from demonstrations with application to manipulation of deformable objects (2014) (27)
- Inverse Reinforcement Learning (2017) (27)
- Max-margin classification of incomplete data (2006) (27)
- An Interaction Design Framework for Social Robots (2012) (27)
- Policy transfer via modularity and reward guiding (2017) (26)
- Arriving on time: estimating travel time distributions on large-scale road networks (2013) (25)
- Combinatorial Energy Learning for Image Segmentation (2015) (25)
- Reset-Free Lifelong Learning with Skill-Space Planning (2020) (25)
- AvE: Assistance via Empowerment (2020) (25)
- CIC: Contrastive Intrinsic Control for Unsupervised Skill Discovery (2022) (25)
- Locally Masked Convolution for Autoregressive Models (2020) (24)
- Learning latent state representation for speeding up exploration (2019) (23)
- Monte Carlo Pose Estimation with Quaternion Kernels and the Bingham Distribution (2012) (22)
- Plan2Vec: Unsupervised Representation Learning by Latent Plans (2020) (22)
- B-Pref: Benchmarking Preference-Based Reinforcement Learning (2021) (22)
- The Importance of Sampling inMeta-Reinforcement Learning (2018) (22)
- LaND: Learning to Navigate From Disengagements (2020) (22)
- Assisted Teleoperation Strategies for Aggressively Controlling a Robot Arm with 2D Input (2012) (22)
- Planning Curvature and Torsion Constrained Ribbons in 3D With Application to Intracavitary Brachytherapy (2015) (21)
- Grasping and Fixturing as Submodular Coverage Problems (2011) (21)
- Transfer Learning for Estimating Causal Effects using Neural Networks (2018) (21)
- Generalization in Robotic Manipulation Through The Use of Non-Rigid Registration (2013) (21)
- Range sensor and silhouette fusion for high-quality 3D Scanning (2015) (20)
- Asynchronous Methods for Model-Based Reinforcement Learning (2019) (20)
- An algorithm for computing customized 3D printed implants with curvature constrained channels for enhancing intracavitary brachytherapy radiation delivery (2013) (19)
- A Constraint-Aware Motion Planning Algorithm for Robotic Folding of Clothes (2012) (19)
- Unsupervised Learning of Visual 3D Keypoints for Control (2021) (19)
- Probabilistically safe policy transfer (2017) (19)
- Hierarchically Decoupled Imitation for Morphological Transfer (2020) (18)
- Finite-Time Regional Verification of Stochastic Nonlinear Systems (2012) (18)
- Inverse Reinforcement Learning via Deep Gaussian Process (2015) (18)
- Masked World Models for Visual Control (2022) (18)
- Selectivity Estimation with Deep Likelihood Models (2019) (18)
- A Disposable Haptic Palpation Probe for Locating Subcutaneous Blood Vessels in Robot-Assisted Minimally Invasive Surgery (2015) (18)
- Lying Pose Recognition for Elderly Fall Detection (2012) (18)
- Policy Learning with Continuous Memory States for Partially Observed Robotic Control (2015) (17)
- Tractability of Planning with Loops (2015) (17)
- Planning locally optimal, curvature-constrained trajectories in 3D using sequential convex optimization (2014) (17)
- Hierarchical Few-Shot Imitation with Skill Transition Models (2021) (17)
- The MineRL BASALT Competition on Learning from Human Feedback (2021) (17)
- Behavioral Priors and Dynamics Models: Improving Performance and Domain Transfer in Offline RL (2021) (17)
- Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation (2018) (16)
- Apprenticeship Learning (2010) (16)
- Exploiting Variable Stiffness in Explosive Movement Tasks (2012) (16)
- Adversarial Motion Priors Make Good Substitutes for Complex Reward Functions (2022) (16)
- Mutual Information State Intrinsic Control (2021) (16)
- Trajectory-wise Multiple Choice Learning for Dynamics Generalization in Reinforcement Learning (2020) (16)
- Skill Preferences: Learning to Extract and Execute Robotic Skills from Human Feedback (2021) (16)
- SFV (2018) (16)
- Sequential quadratic programming for task plan optimization (2016) (15)
- Autonomous Helicopter Flight Using Reinforcement Learning (2010) (15)
- Dynamics Generalization via Information Bottleneck in Deep Reinforcement Learning (2020) (15)
- On the influence of ship motion prediction accuracy on motion planning and control of robotic manipulators on seaborne platforms (2010) (15)
- Controlling Wild Bodies Using Linear Temporal Logic (2012) (15)
- Geometry-Aware Neural Rendering (2019) (15)
- SURF: Semi-supervised Reward Learning with Data Augmentation for Feedback-efficient Preference-based Reinforcement Learning (2022) (14)
- Dimensionality Reduction Flows (2019) (14)
- Frozen Pretrained Transformers as Universal Computation Engines (2022) (14)
- Distributed Robot Ensemble Control for Deployment to Multiple Sites (2012) (14)
- Parallel Training of Deep Networks with Local Updates (2020) (14)
- Towards More Generalizable One-shot Visual Imitation Learning (2021) (13)
- A Framework for Push-Grasping in Clutter (2012) (13)
- Learning Factor Graphs in Polynomial Time & Sample Complexity (2005) (13)
- On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning (2022) (13)
- Benefits of Assistance over Reward Learning (2020) (13)
- A non-rigid point and normal registration algorithm with applications to learning from demonstrations (2015) (13)
- Immunological Computation (2010) (13)
- Task-Agnostic Morphology Evolution (2021) (12)
- Reinforcement Learning with Latent Flow (2021) (12)
- An Empirical Investigation of Representation Learning for Imitation (2022) (12)
- Deep Hierarchical Planning from Pixels (2022) (12)
- Real-Time Prioritized Kinematic Control under Inequality Constraints for Redundant Manipulators (2012) (12)
- Identification and Representation of Homotopy Classes of Trajectories for Search-Based Path Planning in 3D (2012) (12)
- Interpretable and Pedagogical Examples (2017) (11)
- Alpha-Beta Divergences Discover Micro and Macro Structures in Data (2015) (11)
- Mutual Information Maximization for Robust Plannable Representations (2019) (11)
- Cyborg beetles: The remote radio control of insect flight (2010) (11)
- Visual Segmentation of “Simple” Objects for Robots (2012) (11)
- Collision-Free and Curvature-Continuous Path Smoothing in Cluttered Environments (2012) (11)
- Optimized color models for high-quality 3D scanning (2015) (10)
- Unsupervised Surgical Task Segmentation with Milestone Learning (2015) (10)
- Adaptive Online Planning for Continual Lifelong Learning (2019) (10)
- Enabling robots to communicate their objectives (2018) (10)
- Compositional Plan Vectors (2019) (10)
- Dichotomy of Control: Separating What You Can Control from What You Cannot (2022) (10)
- AMP (2020) (10)
- Improving Computational Efficiency in Visual Reinforcement Learning via Stored Embeddings (2021) (9)
- Late Breaking Results: Analog Circuit Generator based on Deep Neural Network enhanced Combinatorial optimization (2019) (9)
- Sliding Manipulation of Rigid Bodies on a Controlled 6-DoF Plate (2012) (9)
- Hindsight Task Relabelling: Experience Replay for Sparse Reward Meta-RL (2021) (8)
- AMP (2021) (8)
- A Learning-Based Approach to Safety for Uncertain Robotic Systems (2018) (8)
- HARP: Autoregressive Latent Video Prediction with High-Fidelity Image Generator (2022) (8)
- Attribute Selection (2010) (8)
- Aligning Text-to-Image Models using Human Feedback (2023) (8)
- Addressing Sample Complexity in Visual Tasks Using HER and Hallucinatory GANs (2019) (8)
- Robotics: Science and Systems VII (2012) (8)
- Learning Universal Policies via Text-Guided Video Generation (2023) (8)
- Playful Interactions for Representation Learning (2021) (8)
- Reward Uncertainty for Exploration in Preference-based Reinforcement Learning (2022) (8)
- TERMES: An Autonomous Robotic System for Three-Dimensional Collective Construction (2011) (7)
- Designing Petri Net Supervisors from LTL Specifications (2012) (7)
- Efficient Online Estimation of Empowerment for Reinforcement Learning (2020) (7)
- Large-Scale Online Expectation Maximization with Spark Streaming (2012) (7)
- Operational Space Control of Constrained and Underactuated Systems (2012) (7)
- VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models (2022) (7)
- The Berkeley 3D Object Dataset (2012) (7)
- Coarse-to-Fine Q-attention with Learned Path Ranking (2022) (6)
- It Takes Four to Tango: Multiagent Selfplay for Automatic Curriculum Generation (2022) (6)
- Plan Arithmetic: Compositional Plan Vectors for Multi-Task Control (2019) (6)
- Budgeted Multi-Armed Bandit Models for Sample-Based Grasp Planning in the Presence of Uncertainty (2014) (6)
- An Art Gallery Approach to Ensuring That Landmarks Are Distinguishable (2012) (6)
- Short – Term Traffic Forecasting : Modeling and Learning Spatio – Temporal Relations in Transportation Networks Using Graph Neural Networks by Behrooz Shahsavari Research Project (2015) (6)
- Adaptive System (2010) (6)
- YCB Benchmarking Project : Object Set, Data Set and Their Applications (特集 ロボットマニピュレーション革命) (2017) (6)
- Efficient Empowerment Estimation for Unsupervised Stabilization (2020) (6)
- Multi-Arm Bandit Models for 2 D Sample Based Grasp Planning with Uncertainty (2015) (6)
- Variable Skipping for Autoregressive Range Density Estimation (2020) (5)
- Fitting geometric programming models to data (2015) (5)
- Masked Autoencoding for Scalable and Generalizable Decision Making (2022) (5)
- Learning to Control a Low-Cost Manipulator Using Data-Efficient Reinforcement Learning (2012) (5)
- Chain of Thought Imitation with Procedure Cloning (2022) (5)
- Beyond lowest-warping cost action selection in trajectory transfer (2015) (5)
- Chain of Hindsight Aligns Language Models with Feedback (2023) (5)
- Large Scale Estimation in Cyberphysical Systems using Streaming Data: a Case Study with Smartphone Traces (2012) (5)
- Sim-to-Real 6D Object Pose Estimation via Iterative Self-training for Robotic Bin-picking (2022) (5)
- Spatial Formation Model for Initiating Conversation (2012) (5)
- Pretraining for Language-Conditioned Imitation with Transformers (2021) (5)
- Foundation Models for Decision Making: Problems, Methods, and Opportunities (2023) (5)
- Blue Gripper: A Robust, Low-Cost, and Force-Controlled Robot Hand (2019) (5)
- Robotics: Science and Systems VIII (2012) (5)
- Cooperative Occlusion-Aware Multi-Robot Target Tracking using Optimization (2015) (5)
- Time Complexity of Sensor-Based Vehicle Routing (2005) (5)
- Learning Efficient Representation for Intrinsic Motivation (2019) (5)
- Likelihood Contribution based Multi-scale Architecture for Generative Flows (2019) (4)
- Coarse-to-fine Q-attention with Tree Expansion (2022) (4)
- JUMBO: Scalable Multi-task Bayesian Optimization using Offline Data (2021) (4)
- Occlusion-aware multi-robot 3D tracking (2016) (4)
- Target Entropy Annealing for Discrete Soft Actor-Critic (2021) (4)
- Scenic4RL: Programmatic Modeling and Generation of Reinforcement Learning Environments (2021) (4)
- Inferring Reward Functions from Demonstrators with Unknown Biases (2018) (4)
- Addressing Distribution Shift in Online Reinforcement Learning with Offline Datasets (2021) (4)
- The Motion Grammar: Linguistic Perception, Planning, and Control (2012) (4)
- Optimization-Based Artifact Correction for Electron Microscopy Image Stacks (2014) (4)
- Bingham Policy Parameterization for 3D Rotations in Reinforcement Learning (2022) (4)
- UCB and InfoGain Exploration via $\boldsymbol{Q}$-Ensembles (2017) (4)
- ManiSkill: Learning-from-Demonstrations Benchmark for Generalizable Manipulation Skills (2021) (3)
- Learning Visual Robotic Control Efficiently with Contrastive Pre-training and Data Augmentation (2020) (3)
- Gaussian Belief Space Planning for Imprecise Articulated Robots (2013) (3)
- One-Shot Composition of Vision-Based Skills from Demonstration (2019) (3)
- Understanding Slip Perception of Soft Fingertips by Modeling and Simulating Stick-Slip Phenomenon (2012) (3)
- Machine Learning for Robotics (2012) (3)
- Fast Wind Turbine Design via Geometric Programming (2013) (3)
- Reducing Variance in Temporal-Difference Value Estimation via Ensemble of Deep Networks (2022) (3)
- Fast Trajectory Correction for Nonholonomic Mobile Robots Using Affine Transformations (2012) (3)
- Preventing Imitation Learning with Adversarial Policy Ensembles (2020) (3)
- Visual Hindsight Experience Replay (2019) (3)
- Robust and Versatile Bipedal Jumping Control through Multi-Task Reinforcement Learning (2023) (3)
- Learning Dynamic Manipulation Skills under Unknown Dynamics with Guided Policy Search (2014) (3)
- Leveraging Critical States to Develop Trust (2017) (3)
- Instruction-Following Agents with Jointly Pre-Trained Vision-Language Models (2022) (2)
- FCM: Forgetful Causal Masking Makes Causal Language Models Better Zero-Shot Learners (2022) (2)
- Guiding Pretraining in Reinforcement Learning with Large Language Models (2023) (2)
- Hierarchical Variational Imitation Learning of Control Programs (2019) (2)
- Temporal-Difference Value Estimation via Uncertainty-Guided Soft Updates (2021) (2)
- Anytime Algorithm (2010) (2)
- A Serial Approach to Handling High-Dimensional Measurements in the Sigma-Point Kalman Filter (2012) (2)
- Fleet-DAgger: Interactive Robot Fleet Learning with Scalable Human Supervision (2022) (2)
- AdaCat: Adaptive Categorical Discretization for Autoregressive Models (2022) (2)
- Spending Thinking Time Wisely: Accelerating MCTS with Virtual Expansions (2022) (2)
- Addressing Sample Complexity in Visual Tasks Using Hindsight Experience Replay and Hallucinatory GANs (2019) (2)
- Patch-based Object-centric Transformers for Efficient Video Generation (2022) (2)
- Portable GPS Baseband Logging (2007) (2)
- Idiot's Bayes (2010) (2)
- Hierarchical Deep Reinforcement Learning Agent with Counter Self-play on Competitive Games (2018) (2)
- Predictive Coding for Boosting Deep Reinforcement Learning with Sparse Rewards (2019) (2)
- Learning What To Do by Simulating the Past (2021) (2)
- Hierarchically Decoupled Morphological Transfer (2020) (2)
- StereoPose: Category-Level 6D Transparent Object Pose Estimation from Stereo Images via Back-View NOCS (2022) (2)
- Load Balancing for Mobility-on-Demand Systems (2012) (2)
- Friendly Patrolling: A Model of Natural Encounters (2012) (2)
- Multi-Level Partitioning and Distribution of the Assignment Problem for Large-Scale Multi-Robot Task Allocation (2012) (2)
- The Wisdom of Hindsight Makes Language Models Better Instruction Followers (2023) (1)
- Explaining Reinforcement Learning Policies through Counterfactual Trajectories (2022) (1)
- Imitating, Fast and Slow: Robust learning from demonstrations via decision-time planning (2022) (1)
- Chapter 5.2 – A Geometric Approach to Robotic Laundry Folding1 (2015) (1)
- CURL: Contrastive Unsupervised Representation Learning for Reinforcement Learning (2020) (1)
- GEM: Group Enhanced Model for Learning Dynamical Control Systems (2021) (1)
- InstructRL: Simple yet Effective Instruction-Following Agents with Multimodal Transformer (2023) (1)
- Geometric Programming for Aircraft Design Optimization Warren Hoburg (2014) (1)
- Tracking-Based Semi-supervised Learning (2012) (1)
- Insect-machine hybrid system (2013) (1)
- From Caging to Grasping (2012) (1)
- Policy Transfer via Modularity (2017) (1)
- Where are we in the search for an Artificial Visual Cortex for Embodied Intelligence? (2023) (1)
- Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling and Design (2022) (1)
- Instruction-Following Agents with Multimodal Transformer (2022) (1)
- GACEM: Generalized Autoregressive Cross Entropy Method for Multi-Modal Black Box Constraint Satisfaction (2020) (1)
- Dynamical System Embedding for Efficient Intrinsically Motivated Artificial Agents (2019) (1)
- Generalization via Information Bottleneck in Deep Reinforcement Learning (2020) (1)
- Programmatic Modeling and Generation of Real-Time Strategic Soccer Environments for Reinforcement Learning (2022) (1)
- Cs294-40 Learning for Robotics and Control Policy Iteration and Function Approximation (2008) (1)
- Count-Based Temperature Scheduling for Maximum Entropy Reinforcement Learning (2021) (1)
- Machine Learning for Helicopter Dynamics Models (2014) (1)
- 6 Relational Markov Networks (2007) (1)
- Learning to Reinforcement Learn by Imitation (2018) (1)
- Learning a System-ID Embedding Space for Domain Specialization with Deep Reinforcement Learning (2018) (1)
- Attribute-Value Learning (2010) (1)
- On Visual Feature Representations for Transition State Learning in Robotic Task Demonstrations (2016) (1)
- Train Offline, Test Online: A Real Robot Learning Benchmark (2022) (1)
- Agent-Based Simulation Models (2010) (1)
- Comparing Heads-Up, Hands-Free Operation of Ground Robots to Teleoperation (2012) (1)
- Automatic Calibration of Multiple Coplanar Sensors (2012) (1)
- Skill-Based Reinforcement Learning with Intrinsic Reward Matching (2022) (1)
- Model-Based Proprioceptive State Estimation for Spring-Mass Running (2012) (1)
- Explaining robot policies (2021) (1)
- Probabilistic Analysis of Correctness of High-Level Robot Behavior with Sensor Error (2012) (1)
- Multi-Environment Pretraining Enables Transfer to Action Limited Datasets (2022) (1)
- Hierarchical Imitation Learning via Variational Inference of Control Programs (2018) (1)
- On Training Robust Policies for Flow Smoothing (2020) (1)
- Teachable Reinforcement Learning via Advice Distillation (2022) (1)
- Responsive Safety in Reinforcement Learning (2020) (1)
- Learning an Interface to Improve Efficiency in Combined Task and Motion Planning (2015) (1)
- Reinforcement Learning Methods to Enable Automatic Tuning of Legged Robots (2012) (1)
- Quantifying Uncertainty in Foundation Models via Ensembles (2022) (1)
- PIXELSNAIL: AN IMPROVED AUTOREGRESSIVE GEN- (2018) (1)
- Tying Surgical Knots From Demonstration : Enhancing Demonstrations and Correcting Errors During Execution (2011) (1)
- Optimization-based Cooperative Multi-Robot Target Tracking with Reasoning about Occlusions (2015) (0)
- Average-Payoff Reinforcement Learning (2010) (0)
- Controllability-Aware Unsupervised Skill Discovery (2023) (0)
- Cyborg Insect : Insect Machine Hybrid System for Locomotion Control (2014) (0)
- Instance Language (2010) (0)
- Kalman Filtering (2018) (0)
- Data-Efficient Exploration with Self Play for Atari (2021) (0)
- Real–time Estimation of Distributed Parameters Systems: Application to Traffic Monitoring (0)
- Unsupervised Reinforcement Learning with Contrastive Intrinsic Control (2022) (0)
- Immune Computing (2010) (0)
- Self-Supervised Learning of Object Motion Through Adversarial Video Prediction (2018) (0)
- Language Quantized AutoEncoders: Towards Unsupervised Text-Image Alignment (2023) (0)
- Average-Cost Neuro-Dynamic Programming (2010) (0)
- PACES TO T RANS-FER S KILLS WITH R EINFORCEMENT L EARNING (2017) (0)
- Masked Trajectory Models for Prediction, Representation, and Control (2023) (0)
- Under review as a conference paper at ICLR 2017 # Exploration : A Study of Count-Based Exploration for Deep Reinforcement Learning (2017) (0)
- Learning Torque-Driven Manipulation Primitives with a Multilayer Neural Network (2014) (0)
- Natural Image Manipulation for Autoregressive Models Using Fisher Scores (2019) (0)
- Algorithms for Multi-task Reinforcement Learning (2020) (0)
- Inverse Entailment (2010) (0)
- Average-Cost Optimization (2010) (0)
- Bottleneck Transformers for Visual Recognition-Supplementary (2021) (0)
- Learning to Select Expert Demonstrations for Deformable Object Manipulation (2010) (0)
- Synkhronos: a Multi-GPU Theano Extension for Data Parallelism (2017) (0)
- PACE N OISE FOR E XPLORATION (2018) (0)
- Data fitting with geometric-programming-compatible softmax functions (2016) (0)
- Biological Flapping and Legged Machine : Insect-Machine Hybrid System (2014) (0)
- Markov Decision Process (2018) (0)
- SELF-SUPERVISED POLICY ADAPTATION (2021) (0)
- Sim-to-Real via Sim-to-Seg: End-to-end Off-road Autonomous Driving Without Real Data (2022) (0)
- Iterative Classification (2010) (0)
- SURF: Semi-supervised Reward Learning with Data Augmentation for Feedback-efficient Preference-based Reinforcement Learning (2022) (0)
- ROTATION PLANE DOUBLY ORTHOGONAL RECUR- (2016) (0)
- Is More Specific Than (2010) (0)
- Autonomous Palpation for Tumor Localization : Design of a Palpation Probe and Gaussian Process Adaptive Sampling (2016) (0)
- Approximate Dynamic Programming (2011) (0)
- Improvement Curve (2010) (0)
- Near Optimal Chernoff Bounds for Markov Decision Processes (2012) (0)
- Universal Planning Networks-Long Version + Supplementary (2018) (0)
- Inverse Optimal Control (2010) (0)
- Inductive Bias (2010) (0)
- Learning Self-Supervised Representations of Code Functionality (2021) (0)
- Innovation Paths for Machine Learning in Robotics [Industry Activities] (2022) (0)
- Bayes Net (2017) (0)
- Languages are Rewards: Chain of Hindsight Finetuning using Human Feedback (2023) (0)
- ON NEURAL NETWORK POLICIES (2017) (0)
- Multi-View Masked World Models for Visual Robotic Manipulation (2023) (0)
- CONSTRAINING INFORMATION FLOW (2018) (0)
- Rotation Plane Doubly Orthogonal Recurrent Neural Networks (2016) (0)
- T ARGET E NTROPY A NNEALING FOR D ISCRETE S OFT A CTOR –C RITIC (2021) (0)
- Analytical Learning (2010) (0)
- Learning 2 D Linear Dynamics in Image Space Using Neural Networks (2014) (0)
- An Open Source AGPS/DGPS Capable C-coded Software Receiver (2009) (0)
- Coupled Dynamical System Based Hand-Arm Grasp Planning under Real-Time Perturbations (2012) (0)
- Autoregressive Uncertainty Modeling for 3D Bounding Box Prediction (2022) (0)
- Induction as Inverted Deduction (2010) (0)
- Deep Learning to Learn (2018) (0)
- Planning to Explore via Latent Disagreement (2020) (0)
- Infinite-Horizon Model Predictive Control for Periodic Tasks with Contacts (2012) (0)
- Imitation Learning of Hierarchical Programs via Variational Inference (2018) (0)
- Inductive Inference Rules (2010) (0)
- Inverse Resolution (2010) (0)
- Absolute Error Loss (2010) (0)
- In-Sample Evaluation (2010) (0)
- A K-fold Method for Baseline Estimation in Policy Gradient Algorithms (2017) (0)
- Immune Network (2010) (0)
- Distributional Instance Segmentation: Modeling Uncertainty and High Confidence Predictions with Latent-MaskRCNN (2023) (0)
- Random Encoder RL Encoder Representation Space Intrinsic Reward Policy Expected Reward Extrinsic Reward (2021) (0)
- Policy Gradient Methods (2008) (0)
- Indirect Reinforcement Learning (2010) (0)
- Instance Space (2010) (0)
- ACTION-DEPENDENT FACTORIZED BASELINES (2018) (0)
- The Prefrontal Cortex and Hierarchical Behavior by Jennifer Sloan A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Neuroscience in the Graduate Division (2013) (0)
- CLUTR: Curriculum Learning via Unsupervised Task Representation Learning (2022) (0)
- Classification (2018) (0)
- An Analytic Motion Planning Solution for the Snakeboard (2012) (0)
- HIRD-P ERSON I MITATION L EARNING (2017) (0)
- Sparse Combinatorial Autoencoders (2015) (0)
- Cross-Entropy Randomized Motion Planning (2012) (0)
- Multi-Objective Policy Gradients with Topological Constraints (2022) (0)
- RoboPianist: A Benchmark for High-Dimensional Robot Control (2023) (0)
- Preference Transformer: Modeling Human Preferences using Transformers for RL (2023) (0)
- Apprenticeship Learning for Robotic Control (2015) (0)
- Bootstrapping Trajectory Transfer from Multiple Demonstrations with Applications to Deformable Object Manipulation (2014) (0)
- Reinforcement Learning for Robotic Assembly with Force Control (2020) (0)
- Towards Better Few-Shot and Finetuning Performance with Forgetful Causal Language Models (2022) (0)
- Associative Bandit Problem (2010) (0)
- Kinematic Cartography for Locomotion at Low Reynolds Numbers (2012) (0)
- Immune-Inspired Computing (2010) (0)
- Is More General Than (2010) (0)
- Title Multi-armed bandit models for 2 D grasp planning with uncertainty (2015) (0)
- Intent Recognition (2010) (0)
- Ballbot : A Low-Cost , Open-Source , Open-Design Robot for Tennis Ball Retrieval (2013) (0)
- Temporally Consistent Video Transformer for Long-Term Video Prediction (2022) (0)
This paper list is powered by the following services:
Other Resources About Pieter Abbeel
What Schools Are Affiliated With Pieter Abbeel?
Pieter Abbeel is affiliated with the following schools: