TITLE. An inverse problem in optimal control.

 

DESCRIPTION OF THE SUBJECT. (See first the proposed subject for “Laurea TriennaleControllo ottimo).  Let us suppose that we have a controlled ordinary differential system in Rn: y’=f(y,u), where u is the control at our disposal, and suppose that we have a control in feedback form u(t)=k(y(t)), that is we have a law, k, that tells us which is the control u to be used with only dependence on our actual position y(t) in Rn. Moreover, let us suppose that the control u should be chosen by a group of agents which, for instance, govern only a “piece” of the evolution y in Rn (for instance only one component), suppose that you are the system manger and that, for some particular reasons, your goal is to force every agents to use the feedback law k. One way to do that is, of course, to say to every agent: “use k!”, but often this is not a good strategy, since the agents are better to remain autonomous. A better strategy seems then to be the use of a sort of “incentive”, that is to find a goal for every agent (for instance to optimize a suitable performance) such that (as you, system manager, has previously checked) the right way to reach the goal is exactly use the law k. Hence, you leave the agents free to choose their control in order to reach the goal, but you know that, if they do the right thing, they use k.

This kind of situation is quite common (and studied) in several frameworks, such as for instance logistic optimization of  distribution of a good through a network of  stocks and stores. However, it seems that a rigorous study with the tools of mathematical analysis and optimal control theory is not yet done.

 

The approach for such a problem is to model the performance criterium (to be devolved to every agent) as the integral over the time of a cost function g depending on the state y(t) and on the control u(t) ,that is ò g(y(t),u(t))dt. The problem is then: given the controlled system y’=f(y,u) and the feedback control u*(t)=k(y(t)), find a function g such that the control u* is optimal for the cost ò g(y(t),u(t))dt, that is it minimizes the cost all over the possible controls. It is clear why this is an “inverse optimal control problem”: usually the control problem is: given the controlled system and the cost, find a control which is optimal. Here instead we already have a control and we want to find a cost in such a way the control turns out to be optimal.

 

 

GOAL OF THE THESIS. Using the Dynamic Programming Principle and the theory of Hamilton-Jacobi equations (which, if not yet known, may be assimilated as first part of the thesis work), try to write down a partial differential equation which, in some suitable sense, must be satisfied by the unknown cost g.

 

REQUIRED NOTIONS. The courses of mathematical analysis, the theory of ordinary differential equations, (suggested but not necessary) some notions on optimal control problems.