The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
The aim of this paper is to propose the technology to allow people to control robots by means of everyday gestures without using sensors or controllers. The hand pose estimation we propose reduces the number of image features per data set to 64, which makes the construction of a large-scale database possible. This has also made it possible to estimate the 3D hand poses of unspecified users with individual...
This paper presents a novel solution to the autonomy of a Portable Robotic Device (PRD) for the visually impaired. The proposed method meets the PRD's requirement of providing 3D navigational information with a small-sized device. The proposed approach is to employ a 3D imaging sensor-the SwissRanger SR4000-for both pose estimation and perception. The SR4000 produces both intensity and range images...
We present the Viewpoint Feature Histogram (VFH), a descriptor for 3D point cloud data that encodes geometry and viewpoint. We demonstrate experimentally on a set of 60 objects captured with stereo cameras that VFH can be used as a distinctive signature, allowing simultaneous recognition of the object and its pose. The pose is accurate enough for robot manipulation, and the computational cost is low...
We present an extension of a neuro-dynamic object recognition system that combines bottom-up recognition of matching patterns and top-down estimation of pose parameters in a recurrent loop. It is extended by an active foveal vision system. Adding the active vision component is easily integrated within the architecture and improves the recognition rate on previous experiments on the COIL-100 database...
The ability to recognize objects and to localize them precisely is essential in all service robotic applications. One of the main challenges for service robots during operation lies in the handling of unavoidable uncertainties which originate from model and sensor inaccuracies and are characteristic for realistic application scenarios. Robustness under real world conditions can only be achieved when...
We present a robotic vision system for object recognition, pose estimation and fast object learning. Our approach uses the Dynamic Neural Field Theory to combine bottom-up recognition of matching patterns and top-down estimation of pose parameters in a recurrent loop. Because Dynamic Neural Fields provide the system with stabilized percepts that still track changes in the incoming sensory stream,...
The latency of a perception system is crucial for a robot performing interactive tasks in dynamic human environments. We present MOPED, a fast and scalable perception system for object recognition and pose estimation. MOPED builds on POSESEQ, a state of the art object recognition algorithm, demonstrating a massive improvement in scalability and latency without sacrificing robustness. We achieve this...
We present an approach for efficiently recognizing all objects in a scene and estimating their full pose from multiple views. Our approach builds upon a state of the art single-view algorithm which recognizes and registers learned metric 3D models using local descriptors. We extend to multiple views using a novel multi-step optimization that processes each view individually and feeds consistent hypotheses...
We propose a combined approach for 3D real-time object recognition and tracking, which is directly applicable to robotic manipulation. We use keypoints features for the initial pose estimation. This pose estimate serves as an initial estimate for edge-based tracking. The combination of these two complementary methods provides an efficient and robust tracking solution. The main contributions of this...
In this paper we address the problem of simultaneous object class and pose estimation using nothing more than object class label measurements from a generic object classifier. We detail a method for designing a likelihood function over the robot configuration space. This function provides a likelihood measure of an object being of a certain class given that the robot (from some position) sees and...
This paper proposes a novel approach to sensor planning for simultaneous object identification and 3D pose estimation. We consider the problem of determining the next-best-view for a movable sensor (or an autonomous agent) to identify an unknown object from among a database of known object models. We use an information theoretic approach to define a metric (based on the difference between the current...
In this paper we present a new approach for labeling 3D points with different geometric surface primitives using a novel feature descriptor - the Fast Point Feature Histograms, and discriminative graphical models. To build informative and robust 3D feature point representations, our descriptors encode the underlying surface geometry around a point p using multi-value histograms. This highly dimensional...
Learning to recognize objects from a small number of example views is a difficult problem of robot vision, of particular importance to assistance robots who are taught by human users. Here we present an approach that combines bottom-up recognition of matching patterns and top-down estimation of pose parameters in a recurrent loop that improves on previous efforts to reconcile invariance of recognition...
This paper proposes a novel global localization approach that uses hybrid maps of objects and spatial layouts. We model indoor environments using the following visual cues from a stereo camera: local invariant features for object recognition and their 3D positions for object location representation. We also use a 2D laser range finder. Therefore, we can build a hybrid local node for a topological...
This paper presents research and development for a hybrid visual servoing control method for a newly developed space robot. The hybrid scheme consists of an eye-in-hand camera and eye-to-hand camera configuration. The first one guarantees good accuracy and the ability to explore the workspace; the second one ensures a panoramic sight of the workspace. The motivation of this paper is to take advantage...
This paper addresses the issue of learning and representing object grasp affordances, i.e. object-gripper relative configurations that lead to successful grasps. The purpose of grasp affordances is to organize and store the whole knowledge that an agent has about the grasping of an object, in order to facilitate reasoning on grasping solutions and their achievability. The affordance representation...
A novel approach is presented which aims at building autonomously visual models of unknown objects, using a humanoid robot. Previous methods have been proposed for the specific problem of the next-best-view during the modeling and the recognition process. However our approach differs as it takes advantage of humanoid specificities in terms of embedded vision sensor and redundant motion capabilities...
Object recognition techniques are well known in the field of machine vision, and aim at the classification of certain observed rigid objects based on the information acquired by a specific sensor. These techniques can either be performed in the 2D image space by simply applying suitable image processing algorithms, or in the real world 3D space by performing surface reconstruction of the object's...
Natural feature image recognition (NFIR) is camera based robotic vision system for recognition, acquisition, tracking and pose estimation of a target vehicle. This paper presents our on-going work on development of the capability of the NFIR software in recognition and acquisition for autonomous rendezvous and docking. The heart of the acquisition algorithm is a classification-based scheme. Feature...
This paper proposes a novel vision-based global localization approach that uses an object and spatial layout based hybrid map. For environment modeling, we use the following visual cues with a stereo camera; local invariant features for object recognition and their 3D positions for object position representation. Also, we use the depth information at the horizontal centerline in images where the optical...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.