Game theoretic controller synthesis for multi-robot motion planning-part II: Policy-based algorithms

Devesh K. Jha, Minghui Zhu, Asok Ray

Research output: Contribution to journalConference article

2 Citations (Scopus)

Abstract

This paper presents the problem of distributed feedback motion planning for multiple robots. The problem of feedback multi-robot motion planning is formulated as a differential noncooperative game. We leverage the existing sampling-based algorithms and value iterations to develop an incremental policy synthesizer. The proposed algorithm makes use of an iterative best response algorithm to incrementally improve the estimate of value functions of the individual robots in the multi-robot motion-planning setting. We show the asymptotic convergence of the limiting policies induced by the proposed Feedback iNash-Policy algorithm for the underlying non-cooperative game. Furthermore, we show that the value iterations allow estimation of the cost-to-go functions for the robots without the requirement on convergence of the value functions for the sampled graph at any particular iteration.

Original languageEnglish (US)
Pages (from-to)168-173
Number of pages6
JournalIFAC-PapersOnLine
Volume28
Issue number22
DOIs
StatePublished - Oct 1 2015
Event5th IFAC Workshop on Distributed Estimation and Control in Networked Systems, NecSys 2015 - Philadelphia, United States
Duration: Sep 10 2015Sep 11 2015

Fingerprint

Motion planning
Robots
Controllers
Feedback
Sampling
Costs

All Science Journal Classification (ASJC) codes

  • Control and Systems Engineering

Cite this

@article{9d92c483f7224328a8988ede1b9a8625,
title = "Game theoretic controller synthesis for multi-robot motion planning-part II: Policy-based algorithms",
abstract = "This paper presents the problem of distributed feedback motion planning for multiple robots. The problem of feedback multi-robot motion planning is formulated as a differential noncooperative game. We leverage the existing sampling-based algorithms and value iterations to develop an incremental policy synthesizer. The proposed algorithm makes use of an iterative best response algorithm to incrementally improve the estimate of value functions of the individual robots in the multi-robot motion-planning setting. We show the asymptotic convergence of the limiting policies induced by the proposed Feedback iNash-Policy algorithm for the underlying non-cooperative game. Furthermore, we show that the value iterations allow estimation of the cost-to-go functions for the robots without the requirement on convergence of the value functions for the sampled graph at any particular iteration.",
author = "Jha, {Devesh K.} and Minghui Zhu and Asok Ray",
year = "2015",
month = "10",
day = "1",
doi = "10.1016/ifacol.2015.10.325",
language = "English (US)",
volume = "28",
pages = "168--173",
journal = "IFAC-PapersOnLine",
issn = "2405-8963",
publisher = "IFAC Secretariat",
number = "22",

}

Game theoretic controller synthesis for multi-robot motion planning-part II : Policy-based algorithms. / Jha, Devesh K.; Zhu, Minghui; Ray, Asok.

In: IFAC-PapersOnLine, Vol. 28, No. 22, 01.10.2015, p. 168-173.

Research output: Contribution to journalConference article

TY - JOUR

T1 - Game theoretic controller synthesis for multi-robot motion planning-part II

T2 - Policy-based algorithms

AU - Jha, Devesh K.

AU - Zhu, Minghui

AU - Ray, Asok

PY - 2015/10/1

Y1 - 2015/10/1

N2 - This paper presents the problem of distributed feedback motion planning for multiple robots. The problem of feedback multi-robot motion planning is formulated as a differential noncooperative game. We leverage the existing sampling-based algorithms and value iterations to develop an incremental policy synthesizer. The proposed algorithm makes use of an iterative best response algorithm to incrementally improve the estimate of value functions of the individual robots in the multi-robot motion-planning setting. We show the asymptotic convergence of the limiting policies induced by the proposed Feedback iNash-Policy algorithm for the underlying non-cooperative game. Furthermore, we show that the value iterations allow estimation of the cost-to-go functions for the robots without the requirement on convergence of the value functions for the sampled graph at any particular iteration.

AB - This paper presents the problem of distributed feedback motion planning for multiple robots. The problem of feedback multi-robot motion planning is formulated as a differential noncooperative game. We leverage the existing sampling-based algorithms and value iterations to develop an incremental policy synthesizer. The proposed algorithm makes use of an iterative best response algorithm to incrementally improve the estimate of value functions of the individual robots in the multi-robot motion-planning setting. We show the asymptotic convergence of the limiting policies induced by the proposed Feedback iNash-Policy algorithm for the underlying non-cooperative game. Furthermore, we show that the value iterations allow estimation of the cost-to-go functions for the robots without the requirement on convergence of the value functions for the sampled graph at any particular iteration.

UR - http://www.scopus.com/inward/record.url?scp=84992491308&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84992491308&partnerID=8YFLogxK

U2 - 10.1016/ifacol.2015.10.325

DO - 10.1016/ifacol.2015.10.325

M3 - Conference article

AN - SCOPUS:84992491308

VL - 28

SP - 168

EP - 173

JO - IFAC-PapersOnLine

JF - IFAC-PapersOnLine

SN - 2405-8963

IS - 22

ER -