Header logo is


2012


An SVD-Based Approach for Ghost Detection and Removal in High Dynamic Range Images
An SVD-Based Approach for Ghost Detection and Removal in High Dynamic Range Images

Srikantha, A., Sidibe, D., Meriaudeau, F.

International Conference on Pattern Recognition (ICPR), pages: 380-383, November 2012 (article)

ps

pdf [BibTex]

2012


pdf [BibTex]


Coupled Action Recognition and Pose Estimation from Multiple Views
Coupled Action Recognition and Pose Estimation from Multiple Views

Yao, A., Gall, J., van Gool, L.

International Journal of Computer Vision, 100(1):16-37, October 2012 (article)

ps

publisher's site code pdf Project Page Project Page Project Page [BibTex]

publisher's site code pdf Project Page Project Page Project Page [BibTex]


{DRAPE: DRessing Any PErson}
DRAPE: DRessing Any PErson

Guan, P., Reiss, L., Hirshberg, D., Weiss, A., Black, M. J.

ACM Trans. on Graphics (Proc. SIGGRAPH), 31(4):35:1-35:10, July 2012 (article)

Abstract
We describe a complete system for animating realistic clothing on synthetic bodies of any shape and pose without manual intervention. The key component of the method is a model of clothing called DRAPE (DRessing Any PErson) that is learned from a physics-based simulation of clothing on bodies of different shapes and poses. The DRAPE model has the desirable property of "factoring" clothing deformations due to body shape from those due to pose variation. This factorization provides an approximation to the physical clothing deformation and greatly simplifies clothing synthesis. Given a parameterized model of the human body with known shape and pose parameters, we describe an algorithm that dresses the body with a garment that is customized to fit and possesses realistic wrinkles. DRAPE can be used to dress static bodies or animated sequences with a learned model of the cloth dynamics. Since the method is fully automated, it is appropriate for dressing large numbers of virtual characters of varying shape. The method is significantly more efficient than physical simulation.

ps

YouTube pdf talk Project Page Project Page [BibTex]

YouTube pdf talk Project Page Project Page [BibTex]


Ghost Detection and Removal for High Dynamic Range Images: Recent Advances
Ghost Detection and Removal for High Dynamic Range Images: Recent Advances

Srikantha, A., Sidib’e, D.

Signal Processing: Image Communication, 27, pages: 650-662, July 2012 (article)

ps

pdf link (url) [BibTex]

pdf link (url) [BibTex]


Visual Servoing on Unknown Objects
Visual Servoing on Unknown Objects

Gratal, X., Romero, J., Bohg, J., Kragic, D.

Mechatronics, 22(4):423-435, Elsevier, June 2012, Visual Servoing \{SI\} (article)

Abstract
We study visual servoing in a framework of detection and grasping of unknown objects. Classically, visual servoing has been used for applications where the object to be servoed on is known to the robot prior to the task execution. In addition, most of the methods concentrate on aligning the robot hand with the object without grasping it. In our work, visual servoing techniques are used as building blocks in a system capable of detecting and grasping unknown objects in natural scenes. We show how different visual servoing techniques facilitate a complete grasping cycle.

am ps

Grasping sequence video Offline calibration video Pdf DOI [BibTex]

Grasping sequence video Offline calibration video Pdf DOI [BibTex]


Visual Orientation and Directional Selectivity Through Thalamic Synchrony
Visual Orientation and Directional Selectivity Through Thalamic Synchrony

Stanley, G., Jin, J., Wang, Y., Desbordes, G., Wang, Q., Black, M., Alonso, J.

Journal of Neuroscience, 32(26):9073-9088, June 2012 (article)

Abstract
Thalamic neurons respond to visual scenes by generating synchronous spike trains on the timescale of 10–20 ms that are very effective at driving cortical targets. Here we demonstrate that this synchronous activity contains unexpectedly rich information about fundamental properties of visual stimuli. We report that the occurrence of synchronous firing of cat thalamic cells with highly overlapping receptive fields is strongly sensitive to the orientation and the direction of motion of the visual stimulus. We show that this stimulus selectivity is robust, remaining relatively unchanged under different contrasts and temporal frequencies (stimulus velocities). A computational analysis based on an integrate-and-fire model of the direct thalamic input to a layer 4 cortical cell reveals a strong correlation between the degree of thalamic synchrony and the nonlinear relationship between cortical membrane potential and the resultant firing rate. Together, these findings suggest a novel population code in the synchronous firing of neurons in the early visual pathway that could serve as the substrate for establishing cortical representations of the visual scene.

ps

preprint publisher's site Project Page [BibTex]

preprint publisher's site Project Page [BibTex]


Bilinear Spatiotemporal Basis Models
Bilinear Spatiotemporal Basis Models

Akhter, I., Simon, T., Khan, S., Matthews, I., Sheikh, Y.

ACM Transactions on Graphics (TOG), 31(2):17, ACM, April 2012 (article)

Abstract
A variety of dynamic objects, such as faces, bodies, and cloth, are represented in computer graphics as a collection of moving spatial landmarks. Spatiotemporal data is inherent in a number of graphics applications including animation, simulation, and object and camera tracking. The principal modes of variation in the spatial geometry of objects are typically modeled using dimensionality reduction techniques, while concurrently, trajectory representations like splines and autoregressive models are widely used to exploit the temporal regularity of deformation. In this article, we present the bilinear spatiotemporal basis as a model that simultaneously exploits spatial and temporal regularity while maintaining the ability to generalize well to new sequences. This factorization allows the use of analytical, predefined functions to represent temporal variation (e.g., B-Splines or the Discrete Cosine Transform) resulting in efficient model representation and estimation. The model can be interpreted as representing the data as a linear combination of spatiotemporal sequences consisting of shape modes oscillating over time at key frequencies. We apply the bilinear model to natural spatiotemporal phenomena, including face, body, and cloth motion data, and compare it in terms of compaction, generalization ability, predictive precision, and efficiency to existing models. We demonstrate the application of the model to a number of graphics tasks including labeling, gap-filling, denoising, and motion touch-up.

ps

pdf project page link (url) [BibTex]

pdf project page link (url) [BibTex]


no image
A new seam-tracking algorithm through characteristic-point detection for a portable welding robot

Chang, D., Son, D., Lee, J., Lee, D., Kim, T., Lee, K., Kim, J.

Robotics and Computer-Integrated Manufacturing, 28, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Two-dimensional autonomous microparticle manipulation strategies for magnetic microrobots in fluidic environments

Pawashe, C., Floyd, S., Diller, E., Sitti, M.

IEEE Transactions on Robotics, 28(2):467-477, IEEE, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


A metric for comparing the anthropomorphic motion capability of artificial hands
A metric for comparing the anthropomorphic motion capability of artificial hands

Feix, T., Romero, J., Ek, C. H., Schmiedmayer, H., Kragic, D.

IEEE RAS Transactions on Robotics, TRO, pages: 974-980, 2012 (article)

ps

Publisher site Human Grasping Database Project [BibTex]

Publisher site Human Grasping Database Project [BibTex]


The Ankyrin 3 (ANK3) Bipolar Disorder Gene Regulates Psychiatric-related Behaviors that are Modulated by Lithium and Stress
The Ankyrin 3 (ANK3) Bipolar Disorder Gene Regulates Psychiatric-related Behaviors that are Modulated by Lithium and Stress

Leussis, M., Berry-Scott, E., Saito, M., Jhuang, H., Haan, G., Alkan, O., Luce, C., Madison, J., Sklar, P., Serre, T., Root, D., Petryshen, T.

Biological Psychiatry , 2012 (article)

ps

Prepublication Article Abstract [BibTex]

Prepublication Article Abstract [BibTex]


no image
Three-dimensional microfiber devices that mimic physiological environments to probe cell mechanics and signaling

Ruder, W. C., Pratt, E. D., Bakhru, S., Sitti, M., Zappe, S., Cheng, C., Antaki, J. F., LeDuc, P. R.

Lab on a Chip, 12(10):1775-1779, Royal Society of Chemistry, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Active visual search in unknown environments using uncertain semantics

Aydemir, Alper, Pronobis, Andrzej, Jensfelt, Patric, Sj, Kristoffer, Aydemir, Alper, Jensfelt, Patric, Aydemir, A, Jensfelt, P, Aydemir, A, Jensfelt, P, others

Transactions, 1, pages: 2329-2335, IEEE, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Modelling of conductive atomic force microscope probes for scanning tunnelling microscope operation

Ozcan, O, Sitti, M

IET Micro \& Nano Letters, 7(4):329-333, IET, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Shape memory polymer-based flexure stiffness control in a miniature flapping-wing robot

Hines, L., Arabagi, V., Sitti, M.

IEEE Transactions on Robotics, 28(4):987-990, IEEE, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Micro-manipulation using rotational fluid flows induced by remote magnetic micro-manipulators

Ye, Z., Diller, E., Sitti, M.

Journal of Applied Physics, 112(6):064912, AIP, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Remotely addressable magnetic composite micropumps

Diller, E., Miyashita, S., Sitti, M.

Rsc Advances, 2(9):3850-3856, Royal Society of Chemistry, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Shape-Programmable Soft Capsule Robots for Semi-Implantable Drug Delivery

Yim, S., Sitti, M.

Mechatronics, IEEE/ASME Transactions on, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Control of multiple heterogeneous magnetic microrobots in two dimensions on nonspecialized surfaces

Diller, E., Floyd, S., Pawashe, C., Sitti, M.

IEEE Transactions on Robotics, 28(1):172-182, IEEE, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Gecko-Inspired Controllable Adhesive Structures Applied to Micromanipulation

Mengüç, Y., Yang, S. Y., Kim, S., Rogers, J. A., Sitti, M.

Advanced Functional Materials, 22(6):1245-1245, WILEY-VCH Verlag, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Elastomer surfaces with directionally dependent adhesion strength and their use in transfer printing with continuous roll-to-roll applications

Yang, S. Y., Carlson, A., Cheng, H., Yu, Q., Ahmed, N., Wu, J., Kim, S., Sitti, M., Ferreira, P. M., Huang, Y., others,

Advanced Materials, 24(16):2117-2122, WILEY-VCH Verlag, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Effect of retraction speed on adhesion of elastomer fibrillar structures

Abusomwan, U., Sitti, M.

Applied Physics Letters, 101(21):211907, AIP, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


Natural Metrics and Least-Committed Priors for Articulated Tracking
Natural Metrics and Least-Committed Priors for Articulated Tracking

Soren Hauberg, Stefan Sommer, Kim S. Pedersen

Image and Vision Computing, 30(6-7):453-461, Elsevier, 2012 (article)

ps

Publishers site Code PDF [BibTex]

Publishers site Code PDF [BibTex]


no image
Impact and Surface Tension in Water: a Study of Landing Bodies

Shih, B., Laham, L., Lee, K. J., Krasnoff, N., Diller, E., Sitti, M.

Bio-inspired Robotics Final Project, Carnegie Mellon University, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Design and rolling locomotion of a magnetically actuated soft capsule endoscope

Yim, S., Sitti, M.

IEEE Transactions on Robotics, 28(1):183-194, IEEE, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Design and manufacturing of a controllable miniature flapping wing robotic platform

Arabagi, V., Hines, L., Sitti, M.

The International Journal of Robotics Research, 31(6):785-800, SAGE Publications Sage UK: London, England, 2012 (article)

pi

[BibTex]

[BibTex]


no image
Chemotactic steering of bacteria propelled microbeads

Kim, D., Liu, A., Diller, E., Sitti, M.

Biomedical microdevices, 14(6):1009-1017, Springer US, 2012 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Hydrogel microrobots actuated by optically generated vapour bubbles

Hu, W., Ishii, K. S., Fan, Q., Ohta, A. T.

Lab on a Chip, 12(19):3821-3826, Royal Society of Chemistry, 2012 (article)

pi

[BibTex]

[BibTex]


Consumer Depth Cameras for Computer Vision - Research Topics and Applications
Consumer Depth Cameras for Computer Vision - Research Topics and Applications

Fossati, A., Gall, J., Grabner, H., Ren, X., Konolige, K.

Advances in Computer Vision and Pattern Recognition, Springer, 2012 (book)

ps

workshop publisher's site [BibTex]

workshop publisher's site [BibTex]

2007


no image
Learning static Gestalt laws through dynamic experience

Ostrovsky, Y., Wulff, J., Sinha, P.

Journal of Vision, 7(9):315-315, ARVO, June 2007 (article)

Abstract
The Gestalt laws (Wertheimer 1923) are widely regarded as the rules that help us parse the world into objects. However, it is unclear as to how these laws are acquired by an infant's visual system. Classically, these “laws” have been presumed to be innate (Kellman and Spelke 1983). But, more recent work in infant development, showing the protracted time-course over which these grouping principles emerge (e.g., Johnson and Aslin 1995; Craton 1996), suggests that visual experience might play a role in their genesis. Specifically, our studies of patients with late-onset vision (Project Prakash; VSS 2006) and evidence from infant development both point to an early role of common motion cues for object grouping. Here we explore the possibility that the privileged status of motion in the developmental timeline is not happenstance, but rather serves to bootstrap the learning of static Gestalt cues. Our approach involves computational analyses of real-world motion sequences to investigate whether primitive optic flow information is correlated with static figural cues that could eventually come to serve as proxies for grouping in the form of Gestalt principles. We calculated local optic flow maps and then examined how similarity of motion across image patches co-varied with similarity of certain figural properties in static frames. Results indicate that patches with similar motion are much more likely to have similar luminance, color, and orientation as compared to patches with dissimilar motion vectors. This regularity suggests that, in principle, common motion extracted from dynamic visual experience can provide enough information to bootstrap region grouping based on luminance and color and contour continuation mechanisms in static scenes. These observations, coupled with the cited experimental studies, lend credence to the hypothesis that static Gestalt laws might be learned through a bootstrapping process based on early dynamic experience.

ps

link (url) DOI [BibTex]

2007


link (url) DOI [BibTex]


Neuromotor prosthesis development
Neuromotor prosthesis development

Donoghue, J., Hochberg, L., Nurmikko, A., Black, M., Simeral, J., Friehs, G.

Medicine & Health Rhode Island, 90(1):12-15, January 2007 (article)

Abstract
Article describes a neuromotor prosthesis (NMP), in development at Brown University, that records human brain signals, decodes them, and transforms them into movement commands. An NMP is described as a system consisting of a neural interface, a decoding system, and a user interface, also called an effector; a closed-loop system would be completed by a feedback signal from the effector to the brain. The interface is based on neural spiking, a source of information-rich, rapid, complex control signals from the nervous system. The NMP described, named BrainGate, consists of a match-head sized platform with 100 thread-thin electrodes implanted just into the surface of the motor cortex where commands to move the hand emanate. Neural signals are decoded by a rack of computers that displays the resultant output as the motion of a cursor on a computer monitor. While computer cursor motion represents a form of virtual device control, this same command signal could be routed to a device to command motion of paralyzed muscles or the actions of prosthetic limbs. The researchers’ overall goal is the development of a fully implantable, wireless multi-neuron sensor for broad research, neural prosthetic, and human neurodiagnostic applications.

ps

pdf [BibTex]

pdf [BibTex]


no image
Bacterial flagella-based propulsion and on/off motion control of microscale objects

Behkam, B., Sitti, M.

Applied Physics Letters, 90(2):023902, AIP, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Friction of partially embedded vertically aligned carbon nanofibers inside elastomers

Aksak, B., Sitti, M., Cassell, A., Li, J., Meyyappan, M., Callen, P.

Applied Physics Letters, 91(6):061906, AIP, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Enhanced friction of elastomer microfiber adhesives with spatulate tips

Kim, S., Aksak, B., Sitti, M.

Applied Physics Letters, 91(22):221913, AIP, 2007 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


On the spatial statistics of optical flow
On the spatial statistics of optical flow

Roth, S., Black, M. J.

International Journal of Computer Vision, 74(1):33-50, 2007 (article)

Abstract
We present an analysis of the spatial and temporal statistics of "natural" optical flow fields and a novel flow algorithm that exploits their spatial statistics. Training flow fields are constructed using range images of natural scenes and 3D camera motions recovered from hand-held and car-mounted video sequences. A detailed analysis of optical flow statistics in natural scenes is presented and machine learning methods are developed to learn a Markov random field model of optical flow. The prior probability of a flow field is formulated as a Field-of-Experts model that captures the spatial statistics in overlapping patches and is trained using contrastive divergence. This new optical flow prior is compared with previous robust priors and is incorporated into a recent, accurate algorithm for dense optical flow computation. Experiments with natural and synthetic sequences illustrate how the learned optical flow prior quantitatively improves flow accuracy and how it captures the rich spatial structure found in natural scene motion.

ps

pdf preprint pdf from publisher [BibTex]

pdf preprint pdf from publisher [BibTex]


no image
Microscale and nanoscale robotics systems [grand challenges of robotics]

Sitti, M.

IEEE Robotics \& Automation Magazine, 14(1):53-60, IEEE, 2007 (article)

pi

[BibTex]

[BibTex]


no image
A new biomimetic adhesive for therapeutic capsule endoscope applications in the gastrointestinal tract

Glass, P., Sitti, M., Appasamy, R.

Gastrointestinal Endoscopy, 65(5):AB91, Mosby, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Visual servoing-based autonomous 2-D manipulation of microparticles using a nanoprobe

Onal, C. D., Sitti, M.

IEEE Transactions on control systems technology, 15(5):842-852, IEEE, 2007 (article)

pi

[BibTex]

[BibTex]


Assistive technology and robotic control using {MI} ensemble-based neural interface systems in humans with tetraplegia
Assistive technology and robotic control using MI ensemble-based neural interface systems in humans with tetraplegia

Donoghue, J. P., Nurmikko, A., Black, M. J., Hochberg, L.

Journal of Physiology, Special Issue on Brain Computer Interfaces, 579, pages: 603-611, 2007 (article)

Abstract
This review describes the rationale, early stage development, and initial human application of neural interface systems (NISs) for humans with paralysis. NISs are emerging medical devices designed to allowpersonswith paralysis to operate assistive technologies or to reanimatemuscles based upon a command signal that is obtained directly fromthe brain. Such systems require the development of sensors to detect brain signals, decoders to transformneural activity signals into a useful command, and an interface for the user.We review initial pilot trial results of an NIS that is based on an intracortical microelectrode sensor that derives control signals from the motor cortex.We review recent findings showing, first, that neurons engaged by movement intentions persist in motor cortex years after injury or disease to the motor system, and second, that signals derived from motor cortex can be used by persons with paralysis to operate a range of devices. We suggest that, with further development, this form of NIS holds promise as a useful new neurotechnology for those with limited motor function or communication.We also discuss the additional potential for neural sensors to be used in the diagnosis and management of various neurological conditions and as a new way to learn about human brain function.

ps

pdf preprint pdf from publisher DOI [BibTex]

pdf preprint pdf from publisher DOI [BibTex]


no image
Adhesion of biologically inspired vertical and angled polymer microfiber arrays

Aksak, B., Murphy, M. P., Sitti, M.

Langmuir, 23(6):3322-3332, ACS Publications, 2007 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Waalbot: An agile small-scale wall-climbing robot utilizing dry elastomer adhesives

Murphy, M. P., Sitti, M.

IEEE/ASME transactions on Mechatronics, 12(3):330-338, IEEE, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Subfeature patterning of organic and inorganic materials using robotic assembly

Tafazzoli, A., Cheng, C., Pawashe, C., Sabo, E. K., Trofin, L., Sitti, M., LeDuc, P. R.

Journal of materials research, 22(06):1601-1608, Cambridge University Press, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Effect of backing layer thickness on adhesion of single-level elastomer fiber arrays

Kim, S., Sitti, M., Hui, C., Long, R., Jagota, A.

Applied Physics Letters, 91(16):161905, AIP, 2007 (article)

pi

[BibTex]

[BibTex]


no image
Adhesion and anisotropic friction enhancements of angled heterogeneous micro-fiber arrays with spherical and spatula tips

Murphy, M. P., Aksak, B., Sitti, M.

Journal of Adhesion Science and Technology, 21(12-13):1281-1296, Taylor & Francis Group, 2007 (article)

pi

Project Page [BibTex]

Project Page [BibTex]


no image
Surface-tension-driven biologically inspired water strider robots: Theory and experiments

Song, Y. S., Sitti, M.

IEEE Transactions on robotics, 23(3):578-589, IEEE, 2007 (article)

pi

[BibTex]

[BibTex]

1998


Summarization of video-taped presentations: Automatic analysis of motion and gesture
Summarization of video-taped presentations: Automatic analysis of motion and gesture

Ju, S. X., Black, M. J., Minneman, S., Kimber, D.

IEEE Trans. on Circuits and Systems for Video Technology, 8(5):686-696, September 1998 (article)

Abstract
This paper presents an automatic system for analyzing and annotating video sequences of technical talks. Our method uses a robust motion estimation technique to detect key frames and segment the video sequence into subsequences containing a single overhead slide. The subsequences are stabilized to remove motion that occurs when the speaker adjusts their slides. Any changes remaining between frames in the stabilized sequences may be due to speaker gestures such as pointing or writing, and we use active contours to automatically track these potential gestures. Given the constrained domain, we define a simple set of actions that can be recognized based on the active contour shape and motion. The recognized actions provide an annotation of the sequence that can be used to access a condensed version of the talk from a Web page.

ps

pdf pdf from publisher DOI [BibTex]

1998


pdf pdf from publisher DOI [BibTex]


Robust anisotropic diffusion
Robust anisotropic diffusion

Black, M. J., Sapiro, G., Marimont, D., Heeger, D.

IEEE Transactions on Image Processing, 7(3):421-432, March 1998 (article)

Abstract
Relations between anisotropic diffusion and robust statistics are described in this paper. Specifically, we show that anisotropic diffusion can be seen as a robust estimation procedure that estimates a piecewise smooth image from a noisy input image. The edge-stopping; function in the anisotropic diffusion equation is closely related to the error norm and influence function in the robust estimation framework. This connection leads to a new edge-stopping; function based on Tukey's biweight robust estimator that preserves sharper boundaries than previous formulations and improves the automatic stopping of the diffusion. The robust statistical interpretation also provides a means for detecting the boundaries (edges) between the piecewise smooth regions in an image that has been smoothed with anisotropic diffusion. Additionally, we derive a relationship between anisotropic diffusion and regularization with line processes. Adding constraints on the spatial organization of the line processes allows us to develop new anisotropic diffusion equations that result in a qualitative improvement in the continuity of edges

ps

pdf pdf from publisher [BibTex]

pdf pdf from publisher [BibTex]


no image
Tele-nanorobotics using an atomic force microscope as a nanorobot and sensor

Sitti, M., Hashimoto, H.

Advanced Robotics, 13(4):417-436, Taylor & Francis, 1998 (article)

pi

[BibTex]

[BibTex]


{PLAYBOT}: A visually-guided robot for physically disabled children
PLAYBOT: A visually-guided robot for physically disabled children

Tsotsos, J. K., Verghese, G., Dickinson, S., Jenkin, M., Jepson, A., Milios, E., Nuflo, F., Stevenson, S., Black, M., Metaxas, D., Culhane, S., Ye, Y., Mann, R.

Image & Vision Computing, Special Issue on Vision for the Disabled, 16(4):275-292, 1998 (article)

Abstract
This paper overviews the PLAYBOT project, a long-term, large-scale research program whose goal is to provide a directable robot which may enable physically disabled children to access and manipulate toys. This domain is the first test domain, but there is nothing inherent in the design of PLAYBOT that prohibits its extension to other tasks. The research is guided by several important goals: vision is the primary sensor; vision is task directed; the robot must be able to visually search its environment; object and event recognition are basic capabilities; environments must be natural and dynamic; users and environments are assumed to be unpredictable; task direction and reactivity must be smoothly integrated; and safety is of high importance. The emphasis of the research has been on vision for the robot this is the most challenging research aspect and the major bottleneck to the development of intelligent robots. Since the control framework is behavior-based, the visual capabilities of PLAYBOT are described in terms of visual behaviors. Many of the components of PLAYBOT are briefly described and several examples of implemented sub-systems are shown. The paper concludes with a description of the current overall system implementation, and a complete example of PLAYBOT performing a simple task.

ps

pdf pdf from publisher DOI [BibTex]

pdf pdf from publisher DOI [BibTex]


EigenTracking: Robust matching and tracking of articulated objects using a view-based representation
EigenTracking: Robust matching and tracking of articulated objects using a view-based representation

Black, M. J., Jepson, A.

International Journal of Computer Vision, 26(1):63-84, 1998 (article)

Abstract
This paper describes an approach for tracking rigid and articulated objects using a view-based representation. The approach builds on and extends work on eigenspace representations, robust estimation techniques, and parameterized optical flow estimation. First, we note that the least-squares image reconstruction of standard eigenspace techniques has a number of problems and we reformulate the reconstruction problem as one of robust estimation. Second we define a “subspace constancy assumption” that allows us to exploit techniques for parameterized optical flow estimation to simultaneously solve for the view of an object and the affine transformation between the eigenspace and the image. To account for large affine transformations between the eigenspace and the image we define a multi-scale eigenspace representation and a coarse-to-fine matching strategy. Finally, we use these techniques to track objects over long image sequences in which the objects simultaneously undergo both affine image motions and changes of view. In particular we use this “EigenTracking” technique to track and recognize the gestures of a moving hand.

ps

pdf pdf from publisher video [BibTex]