Hochschule Reutlingen

Projekte

Hier werden verschiedene Projekte vorgestellt, die im Rahmen des RoboCup bearbeitet werden oder fertiggestellt wurden. Zum Teil enthalten sind Projekte der Vorlesung "Image Understanding" sowie des Wahlfachs "Interactive mobile Robots".

Hier finden Sie die Publikationen von Professor Dr. rer. nat Matthias Rätsch.

Ganz aktuell können Sie hier unsere Open Source Programme herunterladen.

SCITOS

QR-Code/OCR control and speech output

Reading QR-code and OCR.
Using text-to-speech (TTS) software for all outputs, like reaching the goal, as well as error messages or the input from QR-code or OCR.

Conceptional video

  • 16SS:
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Implement a number detection for the floor number in the lift.
    Here you find a short video.
    Here you find a summary.
    Here you find the documentation.
  • Use of commercial Ivona TTS
    Here you find a short video.
    Here you find a summary.
    Here you find the documentation.
  • Localication based on QR code and OCR
    Leonie reads QR-codes and interacts with the user.
    Here you find a short video.
    Here you find a summary.
  • Use of open source software Mary TTS
    Hier you find a short video.
    Hier you find a summary.
  • Using OpenCV for OCR
    Here  you find the documentation.

Gesture detection

Using a 3D sensor to detect body gestures.

  • Gesture detection for SCITOS
    With help of KineticSpace (©Professor Wölfel) gestures are learned, detected and recognized and the Scitos reacts to it.
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Theoretical study to the use of KineticSpace (©Professor Wölfel)
    Here you find a summary.
    Here you find the documentation.
  • Waving detection with MSKinect
    Detecting primitive gestures like waving to trigger SCITOS actions, using the MSKinect.
    Here you find a summary.
    Here you find the documentation.
  • Gesture detection for SCITOS
    With help of KineticSpace (©Professor Wölfel) gestures are learned, detected and recognized and the Scitos reacts to it.
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation (for the PW please contact Felix Ostertag)
  • Gesture detection
    Using the software Kinetic Space gestures are learned and can be recognized afterwards.
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation (for the PW please contact Felix Ostertag)

Leonie recognizes people and looks at them.

In project 1 the microphone array from the MS Kinect is used to detect where the sound comes from. Leonie reacts to whistling for example.

In project 2 the FaceVacs SDK from Cognitec is used to find the faces.

Project 1:

Hiere you find a short video.
Here you find a summary.
Here you find the documentation.

Project 2:

Here you find a short video.
Here you find a summary.
Here you find the documentation.

Grundlagen:

Here you find a summary.
Here you find the documentation.

Leonie recognizes persons using the face recognition software FaceVacs SDK from Cognitec.

  • Implementation of IvonaTTs instead of Mary TTS
    Here you find a summary.
    Here you find the documentation.
  • A camera is used to take pictures of persons. These pictures are stored in a database. Then the system is able to identify known persons on a live stream from a camera.
    Here you find a short video.
    Here you find a summary.
    Here you find the documentation.

Trade Show Robot: Robot presents facial recognition and analysis at trade fairs

Using the Cognitec SDK, the basis of a trade show robot was created. In addition to a guide for the installation and use of the SDK an example of the dialogue as well as a first example program was developed.

Here you will find the resulting package.

Here you will find the short summary.

Detection of human emotions and facial expressions.

  • 16WS: Emotion detection with SVM
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • 16SS: Emotion detection with CNN
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Evaluation of Kinect 2 for emotion detection compared to Kinect 1
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Emotion detection with a neural network
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Emotion detection based on MSKinect and faceshift SDK
    Emotion analysis using a Microsoft Kinect Sensor and an C++ application that receives data from faceshift's network stream.
    With faceshift, 48 action units of the face can be tracked and measured instead of 6 with Kinect SDK.
    Here you find a summary.
    Here and here you find a short video.
  • Emotion detection with MSKinect
    Here you find the documentation.
    Here you find a summary.

(COG Drittmittelprojekt)
Based on an active camera system persons are detected and then tracked via tracking-algorithms.

  • 16SS: Compress and merg of existing stuff and demo zoom cam player
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Dual-Cam-System-Tracking
    On a wide-angle camera a person is tracked, while the PTU is used to fokus a high-res camera on the person.
    Here you find a summary.
    Here you find a short video.
  • Single-Cam-System-Tracking
    Tracking of a person, via a PTU mounted camera.
    Here you find a summary.
    Hier you find a short video.
  • PTU-Steuerung
    Controlling the Pan-Tilt-Unit (PTU) with MIRA.
    Here you find a summary.
    Here you find a short video.
  • PTU-Grundlagen
    Here you find the documentation.
    Here you find a summary.

SCITOS is following a person, including turning and driving backwards, maintaining a constant distance .

Conceptional video

  • 16SS:
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • SCITOS approaching a person
    Here you find a summary.
    Here you find the documentation.
  • Simulation with Simulink
    Here you find a short video.
    Here you find a summary.
    Here you find the documentation.

Different projects concerning possible or actual hardware for SCITOS

16WS:
Ricoh 360° + HTC Vive

Here you find a summary.
Here you find a short video.
Here you find the documentation from.

16SS:
SphereCam

Here you find a summary.
Here you find a short video.
Here you find the documentation.

Navigating over multiple levels with the SCITOS, including using the elevator.

Groundwork for automatic navigation over multiple levels, including a guide about using MIRA center for steering the SCITOS.

Here you find a summary.
Here you find the documentation.

NAO

The NAO recognizes persons by their faces.

  • Improvement of the software to recognize multiple persons on one picture at ones.

Here you find a short video.
Here you find a summary.
Here you find the documentation.

  • Using the FaceVACs SDK from Cognitec, faces are learned and can then be used to recognize persons.

Here you find a short video.
Here you find a summary.
Here you find the documentation.

 

Symptoms of old age arise also in robots. In this case, arthritis in the shoulder and mental overloading.

Unfortunatly there is wear and tear in both Naos Kurt and Bert. Shoulder joints cracking and function is threaten to stop working soon. The same happend in the neck gears, and like them, the shoulders should get new replacement parts. These also need to be designed and made from scratch.

While working, the processors in the heads tend to overheat. This should be taken care of by using better cooling fans.

Both of our Naos have a broken neck - finally the operation is successful!

By a fall, both Naos have suffered a broken neck. The plastic gears have not withstood the strain. Through the commitment of RT-Lion Sasha Brown several thousand € were saved on repair costs. Now a machined aluminum gear supports the center the mind of Kurt and Bert!

Becoming acquainted with the graphical programming interface choreographer as well as the simulation environment Webots.

To protect the real Nao robot and to test programs virtually the simulator in the choreographer or the external simulation tool Webots can be used.

Advantages and disadvantages of both options were captured and documented, as well as their application.

Here you find the data package.

Here you find the documentation.

MixedReality

New installation of the Mixed-Reality server, developed by NorthernStars

Here you find a short video.
Here you find a summary.
Here you find the documentation.

Here you find a ready-to-go VM with the server installed.

New bots for the Mixed-Reality table, developed by NorthernStars

16SS:

Here you find a summary.
Here you find the documentation.

MR-SmallTable–„Find the Bot“
Use of pole filters to improve the image capturing.

To improve the detection of the position marker on the MR table a holder was designed for a polarizing filter. The polarizing filter is to filter out the display on the screen, to thereby detect the marker against a black (for cameras) background easily.

Amazingly, the detection rate has not improved, but actually worsened. Probably because light reflections now have a greater impact on the black background. The accuracy of location detection has, however, improved significantly.

Here you find the documentation.

Defective IR interface makes mixed reality a virtual Reality

Because for some time the IR interface, which makes the connection between bots and server was broken, our mixed reality was limited to pure virtual games.


After many attempts, a working prototype has now been finally created and tested!

Sonstiges

Simultaneous localization and mapping

16SS:

Here you find a summary.
Here you find a short video.
Here you find the documentation from.

Comparing different 3D Scan systems

16WS:

Kinect2 evaluation
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

16SS:

Here you find a summary.
Here you find a short video.
Here you find the documentation from.

Projects dealing with feature analysis like attractiveness.

16SS:

Here you find a summary.
Here you find the documentation from.

A foosball table with an autonomously playing goalie.

Instructional video here.

16WS:
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

16SS:
Here you find a summary.
Here you find the documentation from.

15WS:
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

15SS:
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

14WS:
Here you find a summary.
Here you find the documentation from.

13WS:
Here you find a summary.
Here you find the documentation.

A rover with omni wheels as a test plattform for the Pixy Cam

15WS:
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

Here you find collaborations with the robotics lab of Prof. Gruhler

15WS: Kuka robot web presentation
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

15WS: MANZ Delta Robot
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

15SS: Jigsaw Puzzle Robot
Here you find a summary.
Here you find a short video.
Here you find the documentation from.

Detecting a speaking person, using optical features.

Here you find a short video.
Here you find a summary.
Here you find the documentation from.

By combining a speech-recognition tool and Wolfram Alpha, a robot could answer questions competently.

Here you find a short video.
Here you find a summary.
Here you find the documentation.

Detecting hand gestures for establishing non-verbal communication.

  • CNN for Finger Alphabet
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Keyboard simulation and Finger Alphabet
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Counting fingers with a webcam.
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Hand gestures with Leap Motion, simulating a keyboard as an example
    Here you find a summary.
    Here you find a short video.
    Here you find the documentation.
  • Hand gestures with Leap Motion
    Controlling the SCITOS using hand gestures with Leap Motion.
    Here you find a summary.
    Here you find a short video.
  • Leap Motion talk
    Here you find a summary.
    Here you find a short video.

Data collection and statistics for attractiveness of females within the course "computer science engineer" in the mechanical engineering program. Please know: so ein Algorithmus könnte dazu eingesetzt werden, Diskriminierung bei größeren Bewerbungsprozessen mit hauptsächlich weiblichen Bewerbungen zu erkennen.

Here you find a summary. Bitte beachten Sie, dass so ein Algorithmus dazu eingesetzt werden könnte, Diskriminierung bei größeren Bewerbungsprozessen mit hauptsächlich weiblichen Bewerbungen zu erkennen.

Unterlagen finden Sie hier.

Polymorph yourself into someone else, easy, fast and painless.

Here you find a summary.
Here you find the documentation.

Vortrag zum Thema humanoide Robotik.

Hier finden Sie das entstandene Paket.

Hier finden Sie eine Kurzzusammenfassung.

Different projects about machine learning.

16SS:

Here you find a summary.
Here you find the documentation from.

16SS:

Here you find a summary.
Here you find the documentation from.

Chat Icon Chat Icon