Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 15 de 15
Filter
Add more filters










Publication year range
1.
Sensors (Basel) ; 17(12)2017 Nov 29.
Article in English | MEDLINE | ID: mdl-29186053

ABSTRACT

Tactile sensing is an important perception mode for robots, but the existing tactile technologies have multiple limitations. What kind of tactile information robots need, and how to use the information, remain open questions. We believe a soft sensor surface and high-resolution sensing of geometry should be important components of a competent tactile sensor. In this paper, we discuss the development of a vision-based optical tactile sensor, GelSight. Unlike the traditional tactile sensors which measure contact force, GelSight basically measures geometry, with very high spatial resolution. The sensor has a contact surface of soft elastomer, and it directly measures its deformation, both vertical and lateral, which corresponds to the exact object shape and the tension on the contact surface. The contact force, and slip can be inferred from the sensor's deformation as well. Particularly, we focus on the hardware and software that support GelSight's application on robot hands. This paper reviews the development of GelSight, with the emphasis in the sensing principle and sensor design. We introduce the design of the sensor's optical system, the algorithm for shape, force and slip measurement, and the hardware designs and fabrication of different sensor versions. We also show the experimental evaluation on the GelSight's performance on geometry and force measurement. With the high-resolution measurement of shape and contact force, the sensor has successfully assisted multiple robotic tasks, including material perception or recognition and in-hand localization for robot manipulation.

2.
J Vis ; 17(5): 7, 2017 05 01.
Article in English | MEDLINE | ID: mdl-28505665

ABSTRACT

Color vision provides humans and animals with the abilities to discriminate colors based on the wavelength composition of light and to determine the location and identity of objects of interest in cluttered scenes (e.g., ripe fruit among foliage). However, we argue that color vision can inform us about much more than color alone. Since a trichromatic image carries more information about the optical properties of a scene than a monochromatic image does, color can help us recognize complex material qualities. Here we show that human vision uses color statistics of an image for the perception of an ecologically important surface condition (i.e., wetness). Psychophysical experiments showed that overall enhancement of chromatic saturation, combined with a luminance tone change that increases the darkness and glossiness of the image, tended to make dry scenes look wetter. Theoretical analysis along with image analysis of real objects indicated that our image transformation, which we call the wetness enhancing transformation, is consistent with actual optical changes produced by surface wetting. Furthermore, we found that the wetness enhancing transformation operator was more effective for the images with many colors (large hue entropy) than for those with few colors (small hue entropy). The hue entropy may be used to separate surface wetness from other surface states having similar optical properties. While surface wetness and surface color might seem to be independent, there are higher order color statistics that can influence wetness judgments, in accord with the ecological statistics. The present findings indicate that the visual system uses color image statistics in an elegant way to help estimate the complex physical status of a scene.


Subject(s)
Color Perception/physiology , Color , Light , Wettability , Humans , Psychophysics , Surface Properties
3.
J Vis ; 16(3): 34, 2016.
Article in English | MEDLINE | ID: mdl-26913626

ABSTRACT

Humans can often estimate tactile properties of objects from vision alone. For example, during online shopping, we can often infer material properties of clothing from images and judge how the material would feel against our skin. What visual information is important for tactile perception? Previous studies in material perception have focused on measuring surface appearance, such as gloss and roughness, and using verbal reports of material attributes and categories. However, in real life, predicting tactile properties of an object might not require accurate verbal descriptions of its surface attributes or categories. In this paper, we use tactile perception as ground truth to measure visual material perception. Using fabrics as our stimuli, we measure how observers match what they see (photographs of fabric samples) with what they feel (physical fabric samples). The data shows that color has a significant main effect in that removing color significantly reduces accuracy, especially when the images contain 3-D folds. We also find that images of draped fabrics, which revealed 3-D shape information, achieved better matching accuracy than images with flattened fabrics. The data shows a strong interaction between color and folding conditions on matching accuracy, suggesting that, in 3-D folding conditions, the visual system takes advantage of chromatic gradients to infer tactile properties but not in flattened conditions. Together, using a visual-tactile matching task, we show that humans use folding and color information in matching the visual and tactile properties of fabrics.


Subject(s)
Color Perception/physiology , Form Perception/physiology , Textiles , Touch Perception/physiology , Female , Humans , Imaging, Three-Dimensional , Male , Young Adult
4.
J Vis ; 14(9)2014 Aug 13.
Article in English | MEDLINE | ID: mdl-25122216

ABSTRACT

It is easy to visually distinguish a ceramic knife from one made of steel, a leather jacket from one made of denim, and a plush toy from one made of plastic. Most studies of material appearance have focused on the estimation of specific material properties such as albedo or surface gloss, and as a consequence, almost nothing is known about how we recognize material categories like leather or plastic. We have studied judgments of high-level material categories with a diverse set of real-world photographs, and we have shown (Sharan, 2009) that observers can categorize materials reliably and quickly. Performance on our tasks cannot be explained by simple differences in color, surface shape, or texture. Nor can the results be explained by observers merely performing shape-based object recognition. Rather, we argue that fast and accurate material categorization is a distinct, basic ability of the visual system.


Subject(s)
Form Perception/physiology , Pattern Recognition, Visual/physiology , Recognition, Psychology/physiology , Color , Cues , Humans , Psychophysics , Surface Properties
5.
Int J Comput Vis ; 103(3): 348-371, 2013 Jul 01.
Article in English | MEDLINE | ID: mdl-23914070

ABSTRACT

Our world consists not only of objects and scenes but also of materials of various kinds. Being able to recognize the materials that surround us (e.g., plastic, glass, concrete) is important for humans as well as for computer vision systems. Unfortunately, materials have received little attention in the visual recognition literature, and very few computer vision systems have been designed specifically to recognize materials. In this paper, we present a system for recognizing material categories from single images. We propose a set of low and mid-level image features that are based on studies of human material recognition, and we combine these features using an SVM classifier. Our system outperforms a state-of-the-art system [Varma and Zisserman, 2009] on a challenging database of real-world material categories [Sharan et al., 2009]. When the performance of our system is compared directly to that of human observers, humans outperform our system quite easily. However, when we account for the local nature of our image features and the surface properties they measure (e.g., color, texture, local shape), our system rivals human performance. We suggest that future progress in material recognition will come from: (1) a deeper understanding of the role of non-local surface properties (e.g., extended highlights, object identity); and (2) efforts to model such non-local surface properties in images.

6.
J Opt Soc Am A Opt Image Sci Vis ; 25(4): 846-65, 2008 Apr.
Article in English | MEDLINE | ID: mdl-18382484

ABSTRACT

Human observers can distinguish the albedo of real-world surfaces even when the surfaces are viewed in isolation, contrary to the Gelb effect. We sought to measure this ability and to understand the cues that might underlie it. We took photographs of complex surfaces such as stucco and asked observers to judge their diffuse reflectance by comparing them to a physical Munsell scale. Their judgments, while imperfect, were highly correlated with the true reflectance. The judgments were also highly correlated with certain image statistics, such as moment and percentile statistics of the luminance and subband histograms. When we digitally manipulated these statistics in an image, human judgments were correspondingly altered. Moreover, linear combinations of such statistics allow a machine vision system (operating within the constrained world of single surfaces) to estimate albedo with an accuracy similar to that of human observers. Taken together, these results indicate that some simple image statistics have a strong influence on the judgment of surface reflectance.


Subject(s)
Cues , Data Interpretation, Statistical , Image Interpretation, Computer-Assisted/methods , Lighting/methods , Surface Properties , Task Performance and Analysis , Visual Perception/physiology , Computer Simulation , Humans , Models, Biological , Models, Statistical
7.
Nature ; 447(7141): 206-9, 2007 May 10.
Article in English | MEDLINE | ID: mdl-17443193

ABSTRACT

The world is full of surfaces, and by looking at them we can judge their material qualities. Properties such as colour or glossiness can help us decide whether a pancake is cooked, or a patch of pavement is icy. Most studies of surface appearance have emphasized textureless matte surfaces, but real-world surfaces, which may have gloss and complex mesostructure, are now receiving increased attention. Their appearance results from a complex interplay of illumination, reflectance and surface geometry, which are difficult to tease apart given an image. If there were simple image statistics that were diagnostic of surface properties it would be sensible to use them. Here we show that the skewness of the luminance histogram and the skewness of sub-band filter outputs are correlated with surface gloss and inversely correlated with surface albedo (diffuse reflectance). We find evidence that human observers use skewness, or a similar measure of histogram asymmetry, in making judgements about surfaces. When the image of a surface has positively skewed statistics, it tends to appear darker and glossier than a similar surface with lower skewness, and this is true whether the skewness is inherent to the original image or is introduced by digital manipulation. We also find a visual after-effect based on skewness: adaptation to patterns with skewed statistics can alter the apparent lightness and glossiness of surfaces that are subsequently viewed. We suggest that there are neural mechanisms sensitive to skewed statistics, and that their outputs can be used in estimating surface properties.


Subject(s)
Sculpture , Visual Perception/physiology , Color , Darkness , Humans , Light , Models, Neurological , Optics and Photonics , Surface Properties
8.
IEEE Trans Pattern Anal Mach Intell ; 27(9): 1459-72, 2005 Sep.
Article in English | MEDLINE | ID: mdl-16173188

ABSTRACT

Interpreting real-world images requires the ability distinguish the different characteristics of the scene that lead to its final appearance. Two of the most important of these characteristics are the shading and reflectance of each point in the scene. We present an algorithm that uses multiple cues to recover shading and reflectance intrinsic images from a single image. Using both color information and a classifier trained to recognize gray-scale patterns, given the lighting direction, each image derivative is classified as being caused by shading or a change in the surface's reflectance. The classifiers gather local evidence about the surface's form and color, which is then propagated using the Generalized Belief Propagation algorithm. The propagation step disambiguates areas of the image where the correct classification is not clear from local evidence. We use real-world images to demonstrate results and show how each component of the system affects the results.


Subject(s)
Algorithms , Artificial Intelligence , Colorimetry/methods , Image Enhancement/methods , Image Interpretation, Computer-Assisted/methods , Imaging, Three-Dimensional/methods , Pattern Recognition, Automated/methods , Computer Graphics , Information Storage and Retrieval/methods , Numerical Analysis, Computer-Assisted , Signal Processing, Computer-Assisted
9.
Percept Psychophys ; 67(1): 120-8, 2005 Jan.
Article in English | MEDLINE | ID: mdl-15912876

ABSTRACT

In order to determine the reflectance of a surface, it is necessary to discount luminance changes produced by illumination variation, a process that requires the visual system to respond differently to luminance changes that are due to illumination and reflectance. It is known that various cues can be used in this process. By measuring the strength of lightness illusions, we find evidence that straightness is, used as a cue: When a boundary is straight rather than curved, it has a greater tendency to be discounted, as if it were an illumination edge. The strongest illusions occur when a boundary has high contrast and has multiple X-junctions that preserve a consistent contrast ratio.


Subject(s)
Contrast Sensitivity , Optical Illusions , Orientation , Pattern Recognition, Visual , Attention , Discrimination Learning , Humans , Psychophysics
10.
J Vis ; 4(10): 944-54, 2004 Nov 09.
Article in English | MEDLINE | ID: mdl-15595897

ABSTRACT

Form information related to occlusion is needed to correctly interpret image motion. This work describes one of a series of investigations into the form constraints on motion perception. In the present study, we focus specifically on the geometry of the occluding contour, and in particular on whether its influence on motion can be accounted for merely by its effect on perceived occlusion. We used an occluded square moving in a circle, holding the T-junctions at points of occlusion constant while manipulating the occluding contour. We found evidence for two main influences of occluding contour geometry on motion interpretation and occlusion: the convexity of the occluding contour and additional static T-junctions that are formed elsewhere on the occluding contour. Our results suggest that convex occluding contours are more occlusive than concave ones, and that T-junctions along the contour increase or decrease the strength of occlusion depending on their orientation. Motion interpretation is influenced by both factors, but their effect on motion appears to be dominated by interactions occurring at an intermediate "semilocal" scale, which is larger than the scale at which junctions are defined, but smaller than the scale of the whole moving figure. We propose that these computations are related to occlusion but are not identical to the computations that mediate static occlusion judgments.


Subject(s)
Form Perception/physiology , Motion Perception/physiology , Ocular Physiological Phenomena , Humans , Models, Neurological
11.
J Vis ; 4(9): 798-820, 2004 Sep 23.
Article in English | MEDLINE | ID: mdl-15493971

ABSTRACT

Many materials, including leaves, water, plastic, and chrome exhibit specular reflections. It seems reasonable that the visual system can somehow exploit specular reflections to recover three-dimensional (3D) shape. Previous studies (e.g., J. T. Todd & E. Mingolla, 1983; J. F. Norman, J. T. Todd, & G. A. Orban, 2004) have shown that specular reflections aid shape estimation, but the relevant image information has not yet been isolated. Here we explain how specular reflections can provide reliable and accurate constraints on 3D shape. We argue that the visual system can treat specularities somewhat like textures, by using the systematic patterns of distortion across the image of a specular surface to recover 3D shape. However, there is a crucial difference between textures and specularities: In the case of textures, the image compressions depend on the first derivative of the surface depth (i.e., surface orientation), whereas in the case of specularities, the image compressions depend on the second derivative (i.e., surfaces curvatures). We suggest that this difference provides a cue that can help the visual system distinguish between textures and specularities, even when present simultaneously. More importantly, we show that the dependency of specular distortions on the second derivative of the surface leads to distinctive fields of image orientation as the reflected world is warped across the surface. We find that these "orientation fields" are (i) diagnostic of 3D shape, (ii) remain surprisingly stable when the world reflected in the surface is changed, and (iii) can be extracted from the image by populations of simple oriented filters. Thus the use of specular reflections for 3D shape perception is both easier and more reliable than previous computational work would suggest.


Subject(s)
Form Perception/physiology , Imaging, Three-Dimensional , Light , Cues , Humans
12.
J Vis ; 4(9): 821-37, 2004 Sep 28.
Article in English | MEDLINE | ID: mdl-15493972

ABSTRACT

Although studies of vision and graphics often assume simple illumination models, real-world illumination is highly complex, with reflected light incident on a surface from almost every direction. One can capture the illumination from every direction at one point photographically using a spherical illumination map. This work illustrates, through analysis of photographically acquired, high dynamic range illumination maps, that real-world illumination possesses a high degree of statistical regularity. The marginal and joint wavelet coefficient distributions and harmonic spectra of illumination maps resemble those documented in the natural image statistics literature. However, illumination maps differ from typical photographs in that illumination maps are statistically nonstationary and may contain localized light sources that dominate their power spectra. Our work provides a foundation for statistical models of real-world illumination, thereby facilitating the understanding of human material perception, the design of robust computer vision systems, and the rendering of realistic computer graphics imagery.


Subject(s)
Form Perception/physiology , Lighting , Models, Statistical , Contrast Sensitivity/physiology , Humans , Light
13.
J Vis ; 4(7): 552-63, 2004 Jul 02.
Article in English | MEDLINE | ID: mdl-15330701

ABSTRACT

Form, motion, occlusion, and perceptual organization are intimately related. We sought to assess the role of junctions in their interaction. We used stimuli based on a cross moving within an occluding aperture. The two bars of the cross appear to cohere or move separately depending on the context; in accord with prior literature, motion interpretation depends in part on whether the bar endpoints appear to be occluded. To test the importance of junctions in motion interpretation, we explored the effect of changing the junctions generated at the occlusion points in our stimuli, from T-junctions to L-junctions. In some cases, this change had a large effect on perceived motion; in others, it made little difference, suggesting junctions are not the critical variable. Further experiments suggested that what matters is not junctions per se, but whether illusory contours are introduced when the junction category is changed. Our results are consistent with an optimization-based computation that seeks to minimize the presence of illusory contours in the perceptual representation. Although it may be possible to explain our results with interactions between junctions, parsimony favors an explanation in terms of a cost-function operating on layered surface interpretations, with no explicit reference to junctions.


Subject(s)
Form Perception/physiology , Motion Perception/physiology , Visual Pathways/physiology , Humans , Ocular Physiological Phenomena
14.
J Vis ; 3(5): 347-68, 2003.
Article in English | MEDLINE | ID: mdl-12875632

ABSTRACT

Under typical viewing conditions, we find it easy to distinguish between different materials, such as metal, plastic, and paper. Recognizing materials from their surface reflectance properties (such as lightness and gloss) is a nontrivial accomplishment because of confounding effects of illumination. However, if subjects have tacit knowledge of the statistics of illumination encountered in the real world, then it is possible to reject unlikely image interpretations, and thus to estimate surface reflectance even when the precise illumination is unknown. A surface reflectance matching task was used to measure the accuracy of human surface reflectance estimation. The results of the matching task demonstrate that subjects can match surface reflectance properties reliably and accurately in the absence of context, as long as the illumination is realistic. Matching performance declines when the illumination statistics are not representative of the real world. Together these findings suggest that subjects do use stored assumptions about the statistics of real-world illumination to estimate surface reflectance. Systematic manipulations of pixel and wavelet properties of illuminations reveal that the visual system's assumptions about illumination are of intermediate complexity (e.g., presence of edges and bright light sources), rather than of high complexity (e.g., presence of recognizable objects in the environment).


Subject(s)
Lighting , Sensory Thresholds/physiology , Visual Perception/physiology , Contrast Sensitivity/physiology , Humans , Light , Psychophysics
15.
Nat Neurosci ; 5(6): 598-604, 2002 Jun.
Article in English | MEDLINE | ID: mdl-12021763

ABSTRACT

The pattern of local image velocities on the retina encodes important environmental information. Although humans are generally able to extract this information, they can easily be deceived into seeing incorrect velocities. We show that these 'illusions' arise naturally in a system that attempts to estimate local image velocity. We formulated a model of visual motion perception using standard estimation theory, under the assumptions that (i) there is noise in the initial measurements and (ii) slower motions are more likely to occur than faster ones. We found that specific instantiation of such a velocity estimator can account for a wide variety of psychophysical phenomena.


Subject(s)
Models, Neurological , Motion Perception/physiology , Optical Illusions/physiology , Contrast Sensitivity/physiology , Humans , Time Factors
SELECTION OF CITATIONS
SEARCH DETAIL
...