Seongyong Koo (Koosy)Email: seongyong.koo_at_pickit3d.com |
|
10.2018 –
now
Country Director, Pick-it Korea
09.2017 –
09.2018
Senior computer vision engineer, Pick-it N.V, Leuven, Belgium
04.2015 –
08.2017
Senior scientist in Autonomous
Intelligent Systems Group, Universität
Bonn
05.2014 –
03.2015
Postdoctoral researcher in Dynamic
Human-Robot Interaction Group, Technische
Universität München
02.2009 –
02.2014
PhD in Telerobotics
and Control Laboratory, KAIST,
Daejeon, Korea
09.2011 –
02.2013
Visiting Researcher in Cluster
of Excellence Cognition for Technical Systems, Technische
Universität München
07.2009 –
08.2009
Visiting Researcher in Chair
of Automatic Control Engineering, Technische
Universität München
03.2007 –
02.2009
M.S. in Robotics Prgram,
KAIST, Daejeon, Korea
03.2001 –
02.2007
B.S. in School of
Mechanical Engineering, Hanyang
University, Seoul, Korea
2015 in
Uni-Bonn
MA-INF
4308 Lab Vision Systems: Learning Computer Vision on GPUs
MA-INF
4208 Seminar Vision Systems
2014 in
TUM
WS2015/15 Dynamic Human-Robot Interaction: Visual
tracking in human-robot interaction (Guest lecture)
2014 in
KAIST
Human-Robot Interaction: Visual tracking in human-robot
interaction (Guest lecture)
Dmytro Pavlichenko, Germán Martín García, Seongyong Koo, and
Sven Behnke:
KittingBot: A Mobile Manipulation Robot for Collaborative Kitting in
Automotive Logistics
In Proceedings of 15th International Conference on Intelligent
Autonomous Systems (IAS-15),
Baden-Baden, Germany, June 2018.
Diego Rodriguez, Corbin Cogswell, Seongyong Koo, and Sven
Behnke:
Transferring Grasping Skills to Novel Instances by Latent Space
Non-Rigid Registration
In Proceedings of IEEE International Conference on Robotics and
Automation (ICRA), Brisbane,
Australia, May 2018.
Max Schwarz, Christian Lenz, German Martin Garcia, Seongyong Koo,
Arul Selvam Periyasamy,
Michael Schreiber, and Sven Behnke:
Fast Object Learning and Dual-arm Coordination for Cluttered Stowing,
Picking, and Packing
In Proceedings of IEEE International Conference on Robotics and
Automation (ICRA), Brisbane,
Australia, May 2018.
Seongyong Koo, Grzegorz Ficht, Germán Martín García, Dmytro
Pavlichenko, Martin Raak, and Sven Behnke:
Robolink
Feeder:
Reconfigurable Bin-Picking and Feeding with a Lightweight
Cable-Driven
Manipulator
Accepted
for 13th IEEE International Conference on Automation Science and
Engineering (CASE), Xi'an, China, to appear August 2017.
[Video]
Nikita Araslanov, Seongyong Koo, Juergen Gall, and Sven
Behnke:
Efficient Single-View 3D
Co-segmentation Using Shape Similarity and Spatial Part Relations
In: Rosenhahn B., Andres B. (eds) Pattern Recognition. GCPR 2016.
Lecture Notes in Computer Science, vol 9796. Springer, Cham.
Seongyong Koo and Sven Behnke:
Focused
On-line
Visual-Motor Coordination for a Dual-arm Robot
Manipulator
In Proceedings of IEEE International Conference on Robotics and
Automation
(ICRA), pp. 1579-1586, Stockholm, Sweden, May 2016.
[Video] [Source
code]
Shile Li, Seongyong Koo, and Dongheui Lee:
Real-time
and Model-free Object Tracking using Particle Filter with Joint
Color-Spatial Descriptor
In proceedings of 2015 IEEE/RSJ International Conferene of Intelligent
Robots and Systems (IROS 2015) at Hamburg, Germany, Sep. 2015.
Seongyong Koo, Dongheui Lee, and Dong-Soo Kwon:
Unsupervised
object individuation from RGB-D image sequences
In proceedings of 2014 IEEE/RSJ International Conference on
Intelligent Robots and Systems (IROS 2014) at Chicago, IL, USA, Nov.
2014, pp. 4450-4457.
Seongyong Koo, Dongheui Lee, and Dong-Soo Kwon:
Incremental
object learning and robust tracking of multiple objects from RGB-D
point set data
Journal of Visual Communication and Image Representation (JVCI), Vol.
25, No. 1, pp. 108-121, 2014.
Seongyong Koo, Dongheui Lee, and Dong-Soo Kwon:
Multiple
Object Tracking Using an RGB-D Camera by Hierarchical
Spatiotemporal Data Association
In proceedings of 2013 IEEE/RSJ International Conference on
Intelligent Robots and Systems (IROS 2013) at Tokyo, Japan, Nov. 2013,
pp. 1113-1118.
Seongyong Koo and Dong-Soo Kwon:
Multiple
People Tracking from 2D Depth Data by Deterministic Spatiotemporal
Data Association
In proceedings of 2013 IEEE International Symposium on Robot and Human
Interactive Communication (Ro-Man 2013) at Gyeongju, Korea, Aug. 2013,
pp. 656-661.
Seongyong Koo, Dongheui Lee, and Dong-Soo Kwon:
GMM-based
3D Object Representation and Robust Tracking in Unconstructed
Dynamic Environments
In proceedings of 2013 IEEE International Conference on Robotics and
Automation (ICRA 2013) at Karlsruhe, Germany, May 2013, pp. 1106-1113.
Jaewoo Kim, Seong-Yong Koo, and Dong-Soo Kwon:
A hierarchical model of
expressive robot motion for effective gesture editing
In proceedings of 8th International Conference on Ubiquitous Robots
and Ambient Intelligence (URAI 2011) at Incheon, Korea, Nov. 2011, pp.
23-26.
Seong-Yong Koo, Kiru Park, Dong-Soo Kwon:
A Dual-layer User Model
based Cognitive System for User-Adaptive Service Robots
In proceedings of 20th IEEE International Symposium on Robot and Human
Interactive Communication (Ro-Man 2011) at Atlanta, GA, USA, July
2011, pp. 59-64.
Young-Min Kim, Seong-Yong Koo, Jong Gwan Lim and Dong-Soo
Kwon:
A Robust Online Touch
Recognition for Dynamic Human-robot Interaction
IEEE Transactions on Consumer Electronics, Vol. 56, No. 3,
pp.1979-1987, 2010.
Oh-Hun Kwon, Seong-Yong Koo, Young-Geun Kim, Dong-Soo Kwon:
Telepresence Robot System
for English Tutoring
In proceedings of 6th IEEE Workshop on Advanced Robotics and its
Social Impacts (ARSO 2010) at Seoul, Korea, Oct. 2010, pp. 152-155.
Jong-Chan Park, Hyunsoo Song, Seongyong Koo, Young-Min Kim,
Dong-Soo Kwon:
Robot’s Behavior
Expressions according to the Sentence Types and Emotions with
Modification by Personality
In proceedings of 6th IEEE Workshop on Advanced Robotics and its
Social Impacts (ARSO 2010) at Seoul, Korea, Oct. 2010, pp. 105-110.
Hyoung-Rock Kim, Seong-Yong Koo, and Dong-Soo Kwon:
Designing Reactive Emotion
Generation Model for Interactive Robots
In proceedings of 2010 IEEE/RSJ International Conference on
Intelligent Robots and Systems (IROS 2010) at Taipei, Taiwan, Oct.
2010, pp. 2270-2275. [JTCF Novel Technology Paper Award for
Amusement Culture]
Oh-Hun Kwon, Seong-Yong Koo, Nam-Su Yuk, and Dong-Soo Kwon:
Two-Wheel
Mobile Platform with Active Assistant Wheels for Expressive
Service Robot
In proceedings of 6th International Conference on Ubiquitous Robots
and Ambient Intelligence (URAI 2009) at Gwangju, Korea, Oct. 2009, pp.
152-155.
Seung-Chan Kim, Seongyong Koo, and Dong-Soo Kwon:
Midas Touch - Chunking
Information on a Robotic User Interface using Spatial and
Functional Metaphor
In proceedings of 18th IEEE International Symposium on Robot and Human
Interactive Communication (Ro-Man 2009) at Toyama, Japan, Sep. 2009,
pp. 967-972.
Seongyong Koo and Dong-Soo Kwon:
Recognizing Human
Intentional Actions from the Relative Movements between Human and
Robot
In proceedings of 18th IEEE International Symposium on Robot and Human
Interactive Communication (Ro-Man 2009) at Toyama, Japan, Sep. 2009,
pp 939-944.
Seong-yong Koo, Jong-gwan Lim, and Dong-soo Kwon:
Online Touch Behavior
Recognition of Hard-cover Robot Using Temporal Decision Tree
Classifier
In proceedings of 17th IEEE International Symposium on Robot and Human
Interactive Communication (Ro-Man 2008) at Munich, Germany, Aug. 2008,
pp.425-429.
Seung-Chan Kim, Seong-Yong Koo, Seungyon-Seny Lee, and
Dong-Soo Kwon:
Rubby;
Two Degree-of-Freedom Interface with Haptic and Auditory Feedback
In proceedings of 4th International Conference on Enactive Interfaces
at Grenoble, France, Nov. 2007, pp.357-360.
Seong-Yong Koo and Dong-Soo Kwon:
Gomy
: The Baby Bear-Like Robot for Emotion Human-Robot Interaction
In proceedings of 2007 International Conference on Ubiquitous Robots
and Ambient Intelligent (URAI 2007) at Pohang, Korea, Nov. 2007,
pp.603-608.
PhD dissertation:
Self-constructive multiple
object detection and tracking in unstructured dynamic human
environment
2014. 02, Mechanical Engineering, KAIST
Master dissertation:
Human Intention Recognition
from the Relative Movement between Human and Robot for the
Restaurant Service Robot
2009. 02, Robotics Program, KAIST
4.2015 –
now
ALROMA:
Autonomous Active Object Learning Through Robot Manipulation,
funded by DFG Priority Programme "Autonomous Learning", Germany.
EuRoC: European
Robotics Challenges. Challenge
1: Reconfigurable Interactive Manufacturing Cell, funded by
European Union's Seventh Framework Programme.
[Video]
6.2014 –
3.2014
Learning and control of Dynamic Manipulation Tasks from
Human Demonstrations, funded by DFG Project LE 2731/2-1, Germany
2.2013 –
2.2014
Development of a self-improving bidirectional sustainable
HRI technology, Industrial Strategic Technology Development Program,
funded by the Ministry of Knowledge Economy of Korea
2.2010 –
2.2011
Autonomous Agent of Future Robot Computer, funded by
Ministry of Knowledge Economy and Korea Research Council for
Industrial Science & Technology
8.2009 –
2.2010
Telepresence English Tutoring Robot, Co-work with Rastech
corporation, funded by Ministry of knowledge economy of Korea
12.2008 –
8.2011
Robot Motion Editing Toolkit Software Development, 21st
Century Frontier R&D Program funded by Ministry of Knowledge
Economy of Korea
8.2007 –
2.2008
Touch Interface and Touch Pattern Recognition, 21st
Century Frontier R&D Program funded by Ministry of Knowledge
Economy of Korea
6.2007 –
5.2008
Restaurant Service Robot Development, funded by Human
Resource Training Project for Regional innovation of Korea
Haptically generated
apparatus and portable device having haptically generated apparatus
Korea patent (Patent No 10-1201250, 2012)
Haptically generated
apparatus, portable device and recording medium thereof
Korea patent (Patent No 10- 1214853, 2012)
Method for embedding
emotion onto media utilizing emotion of human or robot
Korea patent (Patent No 10- 1097358, 2011)
System for Clear formatting
and extracting file location using user input and method therefor
Korea patent (Patent No 10-1097835, 2011)
Parallel 2 wheel running
gear and the control method that can control posture and location
separatively through variableness style assistance wheel
Korea patent (Patent No 10-1013936, 2011)
The system for recognizing
of user touch pattern using touch sensor and accelerometer sensor
Korea patent (Patent No 10-0995130, 2010)
Handheld apparatus and
system provided with shock-sensing module, and the method of
executing command using the same
Korea patent (Patent No 10-0936485, 2010)
Universität Bonn, Institute for Computer Science, Departments: I, II, III, IV, V, VI