Category Archives: Projects

Masters Scholarships in Deep Learning & Robotics

Applications for this position are now closed

Masters Scholarships in Deep Learning & Robotics
Istanbul Technical University Faculty of Computer and Informatics Engineering

The Project:

  • Neural networks for robot vision & action.
  • Real robots, state-of-the-art vision.
  • Artificial intelligence.

Will suit those interested in:

  • Artificial intelligence.
  • Machine learning.
  • Computer vision.
  • Robotics.


  • Development and training of neural network architectures.
  • Application to simulated and real mobile robots.
  • Deep learning frameworks (Keras, tensorflow) – Python.
  • Robot Operating System (ROS).
  • Computer vision algorithms development & implementation.
  • Obstacle avoidance, navigation & mapping.

To apply or inquire, email

Masters Scholarship 2014

2 Year Masters Scholarship in Sensory Augmentation

This scholarship is now closed
Though enthusiastic individuals are welcome to make contact about it.


Eye4We have a first-stages project about giving people extra senses using new technologies. We will use depth cameras, such as time-of-flight cameras, to provide 3D data that will be delivered as a soundscape.

Our project is well-specified, funded, and we just need the right person to help us work on it.

This will appeal to those interested in:

  • Artificial intelligence and machine learning.

  • Psychology & cybernetics.

  • Assistive technologies.

  • Spatial audio and signal processing.

To apply, email your CV along with a short message summarising your relevant interests and experience, to

Main Responsibilities

  • Develop physical user interface and sensory augmentation prototype with Arduino.

  • Process sensory 3D data in real time with Point Cloud Library.

  • Produce spatial soundscapes using HARK and SLAB3D.


Application Deadline: December 31, 2014.

Estimated Duration: January/February 2015 – January/February 2017.

Scholarship amount: 1500TL/month, not taxed.

Main Requirements: Registered full-time for a Masters degree in Turkey.

Principle Investigator: Damien Jade Duff,

Researcher: Gökhan İnce,



Perceiving while planning

Consider what the eyes are doing when involved in the solving of a jigsaw puzzle. While the mind is darting about, imagining placement possibilities, considering combinations, and pondering strategies, the eyes too are darting from place to place over the puzzle, examining pieces relevant to a considered placement, checking edges for compatibility, and studying the layout. The eyes are responding to the deliberation of the mind, checking expectations and seeking out necessary information. They are assuring that the deliberation is rooted in the physical reality of the problem. [1]

This thought motivates work [1,2] that I presented yesterday at the Knowledge Representation and Reasoning for Robotics workshop at ICLP 2013 in Istanbul (work that results from the current TÜBİTAK-funded leg of my collaboration with Sabancı University Cognitive Robotics laboratory).

There is lots and lots of lovely work on planning for perceiving, planning under uncertainty, and planning in a changing world, topics that are of immediate interest to roboticists. In the current work we take a different angle and, for a while, do away with those considerations and just concentrate on what is the best way of integrating perceptual processing with classical planning, in terms of conceptual simplicity and in terms of efficiency. It is rather preliminary work.

For more information, check out the paper [1] or the slides [2] from yesterday’s talk, where I go into the motivation behind planning for manipulation and why mobile manipulation is a ripe topic; also the graphics are pretty. Unfortunately the movies will only be embedded if you use Libre/Open-Office (there is a PDF version too).

[1] Duff, D. J., Erdem, E., & Patoglu, V. (2013). Integration of 3D Object Recognition and Planning for Robotic Manipulation: A Preliminary Report. ICLP-KRR 2013, Istanbul. Retrieved from

[2] (slides)


Project: Integration of sensing, planning and action

Since September I have been working at the Sabancı Üniversitesi Cognitive Robotics Laboratory (CogRobo Lab) on a new project supported by the TÜBİTAK 2216 international post-doctoral fund. I was attracted to the CogRobo Lab because the work there focuses on the interface between “high-level” and “low-level” reasoning, and I am particularly interested in the relationship between physical and abstract cognition. The CogRobo Lab is also ambitious in terms of the potential applications of the technology we are developing there.

The premise on which the CogRobo Lab was founded is the premise that logic programming is a software engineering tool with a lot of promise in allowing the specification of robot problems and the generation of abstract solutions to them. This promise relies in turn on logic databases’ promise of extensibility (roughly, due to the ability to add new facts to knowledge-bases), and the existence of increasingly powerful general purpose reasoners. Action planning is one area where logic programming is customarily applied in a robotics domain.

Unfortunately, in order to make such an approach work in real systems a lot of work must be done to bring these abstract formalisms into the real-world; they are difficult to apply in practice due to the vast number of “low-level” behaviours or computations necessary to make them work, and frequently these low-level computations are brought in as a range of special-purpose hacks, and when they are not hacks they do tend to remain special-purpose due to the time-consuming nature of their development (and due to the fact that there are many difficult problems still to be solved in this area).

As such the approach at the CogRobo Lab is to try and realise the promise of logic programming by exploring the necessary interfaces between the general purpose reasoning mechanisms and more special purpose ones. The idea is to create general purpose components and interfaces that can support logic programming as a robot programming paradigm. In the area of planning, hybrid logic programming is the name for this area of research and one approach to it particularly explored at present at the Cogrobo Lab is that of the external predicate or semantic attachment.

My own research at the lab is about investigating ways in which sensory processing components can be constructed so that they can be re-used by logic programs, so that they are reflexive to the information needed by planning (so that unnecessary sensory computation is not done, giving more room for more relevant expensive computation), and so that they are able to adapt to and possibly make use of low-bandwidth information supplied through the “high-level” interfaces (for example, verbal knowledge). Furthermore, they must be integrated with geometric, kinematic and dynamic reasoning (for example, motion planning, object stability analysis).

To this end I am investigating different paradigms of data association between sensory and high-level components, different procedural models of integration with planning, as well as reflexive sensory computation. All of this is in the context of mobile manipulation with a Kinect-equipped Kuka YouBot within the domain of the Robocup@Work competition, with difficult constrained stacking (e.g. shelving) problems.

The CogRobo Lab has its web-page here: