Remember Me
Or use your Academic/Social account:


Or use your Academic/Social account:


You have just completed your registration at OpenAire.

Before you can login to the site, you will need to activate your account. An e-mail will be sent to you with the proper instructions.


Please note that this site is currently undergoing Beta testing.
Any new content you create is not guaranteed to be present to the final version of the site upon release.

Thank you for your patience,
OpenAire Dev Team.

Close This Message


Verify Password:
Verify E-mail:
*All Fields Are Required.
Please Verify You Are Human:
fbtwitterlinkedinvimeoflicker grey 14rssslideshare1
Dubba, KSR; De Oliveira, MR; Lim, GH; Kasaei, H; Lopes, LS; Tomé, A; Cohn, AG (2014)
Publisher: AI Access Foundation
Languages: English
Types: Other
In order to behave autonomously, it is desirable for robots to have the ability to use human supervision and learn from different input sources (perception, gestures, verbal and textual descriptions etc). In many machine learning tasks, the supervision is directed specifically towards machines and hence is straight forward clearly annotated examples. But this is not always very practical and recently it was found that the most preferred interface to robots is natural language. Also the supervision might only be available in a rather indirect form, which may be vague and incomplete. This is frequently the case when humans teach other humans since they may assume a particular context and existing world knowledge. We explore this idea here in the setting of conceptualizing objects and scene layouts. Initially the robot undergoes training from a human in recognizing some objects in the world and armed with this acquired knowledge it sets out in the world to explore and learn more higher level concepts like static scene layouts and environment activities. Here it has to exploit its learned knowledge and ground language into perception to use inputs from different sources that might have overlapping as well as novel information. When exploring, we assume that the robot is given visual input, without explicit type labels for objects, and also that it has access to more or less generic linguistic descriptions of scene layout. Thus our task here is to learn the spatial structure of a scene layout and simultaneously visual object models it was not trained on. In this paper, we present a cognitive architecture and learning framework for robot learning through natural human supervision and using multiple input sources by grounding language in perception.
  • The results below are discovered through our pilot algorithms. Let us know how we are doing!

    • Balbiani, P.; Condotta, J. F.; and Del Cerro, L. F. 1998. A model for reasoning about bidimensional temporal relations.
    • In Principles of KRR, 124-130.
    • Bentley, J. L. 1975. Multidimensional binary search trees Kuhn, H. W. 1955. The Hungarian method for the assignment problem. Naval research logistics quarterly 2(1-2):83- 97.
    • Lim, G. H.; Suh, I. H.; and Suh, H. 2011. Ontology-based unified robot knowledge for service robots in indoor environments. Systems, Man and Cybernetics, Part A: Systems and Humans, IEEE Transactions on 41(3):492-509.
    • Matuszek, C.; Fitzgerald, N.; Zettlemoyer, L.; Bo, L.; and Fox, D. 2012. A joint model of language and perception for grounded attribute learning. In ICML, 1671-1678.
    • Mavridis, N., and Roy, D. 2006. Grounded situation models for robots: Where words and percepts meet. In Intelligent Robots and Systems, 2006 International Conference on, 4690-4697.
    • Nau, D. S.; Au, T.-C.; Ilghami, O.; Kuter, U.; Murdock, J. W.; Wu, D.; and Yaman, F. 2003. SHOP2: An HTN planning system. Journal of AI Research 20:379-404.
    • Nedas, K. A., and Egenhofer, M. J. 2008. Spatial-scene similarity queries. Transactions in GIS 12(6):661-681.
    • Noy, N. F., and Musen, M. A. 1999. An algorithm for merging and aligning ontologies: Automation and tool support.
    • In Workshop on Ontology Management at AAAI-1999.
    • Oliveira, M.; Lim, G. H.; Kasaei, H.; Chauhan, A.; Seabra Lopes, L.; and Tome´, A. M. 2014. A software architecture for open-ended object conceptualization from experiences.
    • In ICRA 2014, submitted.
    • Potmesil, M. 1987. Generating octree models of 3D objects from their silhouettes in a sequence of images. Computer Vision, Graphics, and Image Processing 40(1):1-29.
    • Rusu, R., and Cousins, S. 2011. 3D is here: Point Cloud Library (PCL). In ICRA, 2011, 1-4.
    • Scopelliti, M.; Giuliani, M. V.; and Fornara, F. 2005. Robots in a domestic setting: a psychological approach. Universal Access in the Information Society 4(2):146-155.
    • Seabra Lopes, L., and Chauhan, A. 2007. How many words can my robot learn?: An approach and experiments with one-class learning. Interaction Studies 8(1):53-81.
    • Seabra Lopes, L., and Chauhan, A. 2008. Open-ended category learning for language acquisition. Connection Science 20(4):277-297.
    • Tulving, E. 1991. Concepts of human memory. In Memory: Organization and locus of change. Oxford Univ. Press. 3- 32.
  • No related research data.
  • No similar publications.
  • BioEntity Site Name
    Google Code

Share - Bookmark

Funded by projects

  • EC | RACE

Cite this article