Reinforcement Learning and Transfer Learning: Simulation-Robot System For Object-Handling
Reinforcement Learning and Transfer Learning: Simulation-Robot System For Object-Handling
Summary
Introduction
Contents
1
elements, e.g.: energy used, and/or maintenance of components. Considering, then,
the mishaps, recently-done analyses seek to make use of similar and previously
processes data so that the optimal decision can be made more nimbly.
An important paradigm for robots to develop an acceptable control over their
own motors, which will be approached by this project is the Reinforcement Learning
– a topic in the field of Machine Learning – that enables computer programs and
robots to fulfill a certain duty through trial and error. Different tasks can currently be
taught, like: walking, folding shirts, handling objects, among others.
Another sphere of study explored is Transfer Learning. This area of research
is based on the utilization of the expertise obtained after the optimization of a task
to improve another, understanding the second to be of a related domain. Such
proposal supplies an array of benefits when compared to old methods: the reduction
of the processing (that would otherwise make it unfeasible) being the main one.
The linking of these two groups of techniques, used in similar ways in articles
referenced ([1], [2], [3] e [4]), will be implemented to achieve the proposed scientific
investigations. The first one is developing efficient algorithms of Reinforcement
Learning, optimizing the object-handling assignments through the means of self-
learning with the use of vision-based robots (based on activities made in
backgrounds such as the Amazon Picking Challenge). The second inquiry will be
exploring different possibilities for the transference of simulation-acquired
knowledge to an actual robot, with the intention of minimizing the computational cost
in order to learn such task, and, finally, yield better results.
To achieve the proposed goals, this research will perform four steps of
scientific development:
1. Contextualization:
This step will be responsible for keeping scientific coherence, revising the
academic literature, so that it is possible to elaborate a tools and knowledge map
that are in accord with what could and should be used during the development of
the project. Some examples follow: classical techniques of Reinforcement Learning,
2
different architectures of Neural Networks, Neural Networks applied to
Reinforcement Learning and optimizing Evolutionary Algorithms.
Estimated duration: 3 months.
3. Test phase:
Having the environment built, all possible and viable approaches found will
be used, that are within the span of time allotted for the project, and that appear to
yield promising results. It will be possible, then, to test new methods, such as the
Deep Neural Networks, to evaluate the obtained algorithms.
Estimated duration: 6 months.