Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 15 de 15
Filtrar
Más filtros










Base de datos
Intervalo de año de publicación
1.
Behav Res Methods ; 2024 Mar 05.
Artículo en Inglés | MEDLINE | ID: mdl-38443726

RESUMEN

Virtual reality (VR) displays are being used in an increasingly wide range of applications. However, previous work shows that viewers often perceive scene properties very differently in real and virtual environments and so realistic perception of virtual stimuli should always be a carefully tested conclusion, not an assumption. One important property for realistic scene perception is surface color. To evaluate how well virtual platforms support realistic perception of achromatic surface color, we assessed lightness constancy in a physical apparatus with real lights and surfaces, in a commercial VR headset, and on a traditional flat-panel display. We found that lightness constancy was good in all three environments, though significantly better in the real environment than on the flat-panel display. We also found that variability across observers was significantly greater in VR and on the flat-panel display than in the physical environment. We conclude that these discrepancies should be taken into account in applications where realistic perception is critical but also that in many cases VR can be used as a flexible alternative to flat-panel displays and a reasonable proxy for real environments.

2.
J Vis ; 22(1): 7, 2022 01 04.
Artículo en Inglés | MEDLINE | ID: mdl-35024760

RESUMEN

During visual search, attention is guided by specific features, including shape. Our understanding of shape guidance is limited to specific attributes (closures and line terminations) that do not fully explain the richness of preattentive shape processing. We used a novel genetic algorithm method to explore shape space and to stimulate hypotheses about shape guidance. Initially, observers searched for targets among 12 random distractors defined, in radial frequency space, by the amplitude and phase of 10 radial frequencies. Reaction time (RT) was the measure of "fitness." To evolve toward an easier search task, distractors with faster RTs survived to the next generation, "mated," and produced offspring (new distractors for the next generation of search). To evolve a harder search, surviving distractors were those yielding longer RTs. Within eight generations of evolution, the method succeeds in producing visual searches either harder or easier than the starting search. In radial frequency space, easy distractors evolve amplitude × frequency spectra that are dissimilar to the target, whereas hard distractors evolve spectra that are more similar to the target. This method also works with naturally shaped targets (e.g., rabbit silhouettes). Interestingly, the most inefficient distractors featured a combination of a body and ear distractors that did not resemble the rabbit (visually or in spectrum). Adding extra ears to these distractors did not impact the search spectrally and instead made it easier to confirm a rabbit, once it was found. In general, these experiments show that shapes that are clearly distinct when attended are similar to each other preattentively.


Asunto(s)
Algoritmos , Atención , Animales , Reconocimiento Visual de Modelos , Conejos , Tiempo de Reacción
3.
J Vis ; 21(2): 8, 2021 02 03.
Artículo en Inglés | MEDLINE | ID: mdl-33595646

RESUMEN

Categorization performance is a popular metric of scene recognition and understanding in behavioral and computational research. However, categorical constructs and their labels can be somewhat arbitrary. Derived from exhaustive vocabularies of place names (e.g., Deng et al., 2009), or the judgements of small groups of researchers (e.g., Fei-Fei, Iyer, Koch, & Perona, 2007), these categories may not correspond with human-preferred taxonomies. Here, we propose clustering by increasing the rand index via coordinate ascent (CIRCA): an unsupervised, data-driven clustering method for deriving ground-truth scene categories. In Experiment 1, human participants organized 80 stereoscopic images of outdoor scenes from the Southampton-York Natural Scenes (SYNS) dataset (Adams et al., 2016) into discrete categories. In separate tasks, images were grouped according to i) semantic content, ii) three-dimensional spatial structure, or iii) two-dimensional image appearance. Participants provided text labels for each group. Using the CIRCA method, we determined the most representative category structure and then derived category labels for each task/dimension. In Experiment 2, we found that these categories generalized well to a larger set of SYNS images, and new observers. In Experiment 3, we tested the relationship between our category systems and the spatial envelope model (Oliva & Torralba, 2001). Finally, in Experiment 4, we validated CIRCA on a larger, independent dataset of same-different category judgements. The derived category systems outperformed the SUN taxonomy (Xiao, Hays, Ehinger, Oliva, & Torralba, 2010) and an alternative clustering method (Greene, 2019). In summary, we believe this novel categorization method can be applied to a wide range of datasets to derive optimal categorical groupings and labels from psychophysical judgements of stimulus similarity.


Asunto(s)
Reconocimiento Visual de Modelos/fisiología , Reconocimiento en Psicología/fisiología , Adolescente , Adulto , Atención/fisiología , Femenino , Humanos , Juicio , Masculino , Psicofísica , Corteza Visual/fisiología , Adulto Joven
4.
J Med Imaging (Bellingham) ; 4(4): 045501, 2017 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-29098168

RESUMEN

As a promising imaging modality, digital breast tomosynthesis (DBT) leads to better diagnostic performance than traditional full-field digital mammograms (FFDM) alone. DBT allows different planes of the breast to be visualized, reducing occlusion from overlapping tissue. Although DBT is gaining popularity, best practices for search strategies in this medium are unclear. Eye tracking allowed us to describe search patterns adopted by radiologists searching DBT and FFDM images. Eleven radiologists examined eight DBT and FFDM cases. Observers marked suspicious masses with mouse clicks. Eye position was recorded at 1000 Hz and was coregistered with slice/depth plane as the radiologist scrolled through the DBT images, allowing a 3-D representation of eye position. Hit rate for masses was higher for tomography cases than 2-D cases and DBT led to lower false positive rates. However, search duration was much longer for DBT cases than FFDM. DBT was associated with longer fixations but similar saccadic amplitude compared with FFDM. When comparing radiologists' eye movements to a previous study, which tracked eye movements as radiologists read chest CT, we found DBT viewers did not align with previously identified "driller" or "scanner" strategies, although their search strategy most closely aligns with a type of vigorous drilling strategy.

5.
J Vis ; 16(2): 13, 2016 Jan 01.
Artículo en Inglés | MEDLINE | ID: mdl-27893077

RESUMEN

People are good at rapidly extracting the "gist" of a scene at a glance, meaning with a single fixation. It is generally presumed that this performance cannot be mediated by the same encoding that underlies tasks such as visual search, for which researchers have suggested that selective attention may be necessary to bind features from multiple preattentively computed feature maps. This has led to the suggestion that scenes might be special, perhaps utilizing an unlimited capacity channel, perhaps due to brain regions dedicated to this processing. Here we test whether a single encoding might instead underlie all of these tasks. In our study, participants performed various navigation-relevant scene perception tasks while fixating photographs of outdoor scenes. Participants answered questions about scene category, spatial layout, geographic location, or the presence of objects. We then asked whether an encoding model previously shown to predict performance in crowded object recognition and visual search might also underlie the performance on those tasks. We show that this model does a reasonably good job of predicting performance on these scene tasks, suggesting that scene tasks may not be so special; they may rely on the same underlying encoding as search and crowded object recognition. We also demonstrate that a number of alternative "models" of the information available in the periphery also do a reasonable job of predicting performance at the scene tasks, suggesting that scene tasks alone may not be ideal for distinguishing between models.


Asunto(s)
Fijación Ocular/fisiología , Reconocimiento Visual de Modelos/fisiología , Campos Visuales/fisiología , Adulto , Femenino , Humanos , Masculino , Persona de Mediana Edad , Encuestas y Cuestionarios , Adulto Joven
6.
Atten Percept Psychophys ; 78(7): 2135-51, 2016 10.
Artículo en Inglés | MEDLINE | ID: mdl-27192994

RESUMEN

Suppose that you are looking for visual targets in a set of images, each containing an unknown number of targets. How do you perform that search, and how do you decide when to move from the current image to the next? Optimal foraging theory predicts that foragers should leave the current image when the expected value from staying falls below the expected value from leaving. Here, we describe how to apply these models to more complex tasks, like search for objects in natural scenes where people have prior beliefs about the number and locations of targets in each image, and search is guided by target features and scene context. We model these factors in a guided search task and predict the optimal time to quit search. The data come from a satellite image search task. Participants searched for small gas stations in large satellite images. We model quitting times with a Bayesian model that incorporates prior beliefs about the number of targets in each map, average search efficiency (guidance), and actual search history in the image. Clicks deploying local magnification were used as surrogates for deployments of attention and, thus, for time. Leaving times (measured in mouse clicks) were well-predicted by the model. People terminated search when their expected rate of target collection fell to the average rate for the task. Apparently, people follow a rate-optimizing strategy in this task and use both their prior knowledge and search history in the image to decide when to quit searching.


Asunto(s)
Atención , Toma de Decisiones , Procesamiento Espacial , Percepción Visual , Teorema de Bayes , Humanos , Modelos Psicológicos , Reconocimiento Visual de Modelos , Factores de Tiempo
7.
Atten Percept Psychophys ; 78(4): 978-87, 2016 05.
Artículo en Inglés | MEDLINE | ID: mdl-26846753

RESUMEN

Previous work has shown that human observers discount or neglect cast shadows in natural and artificial scenes across a range of visual tasks. This is a reasonable strategy for a visual system designed to recognize objects under a range of lighting conditions, since cast shadows are not intrinsic properties of the scene-they look different (or disappear entirely) under different lighting conditions. However, cast shadows can convey useful information about the three-dimensional shapes of objects and their spatial relations. In this study, we investigated how well people detect changes to cast shadows, presented in natural scenes in a change blindness paradigm, and whether shadow changes that imply the movement or disappearance of an object are more easily noticed than shadow changes that imply a change in lighting. In Experiment 1, a critical object's shadow was removed, rotated to another direction, or shifted down to suggest that the object was floating. All of these shadow changes were noticed less often than changes to physical objects or surfaces in the scene, and there was no difference in the detection rates for the three types of changes. In Experiment 2, the shadows of visible or occluded objects were removed from the scenes. Although removing the cast shadow of an occluded object could be seen as an object deletion, both types of shadow changes were noticed less often than deletions of the visible, physical objects in the scene. These results show that even informative shadow changes are missed, suggesting that cast shadows are discounted fairly early in the processing of natural scenes.


Asunto(s)
Percepción de Profundidad , Iluminación , Estimulación Luminosa/métodos , Percepción Visual , Adulto , Ceguera , Humanos , Movimiento
8.
Behav Res Methods ; 48(4): 1343-1348, 2016 12.
Artículo en Inglés | MEDLINE | ID: mdl-26660198

RESUMEN

Change blindness has been a topic of interest in cognitive sciences for decades. Change detection experiments are frequently used for studying various research topics such as attention and perception. However, creating change detection stimuli is tedious and there is no open repository of such stimuli using natural scenes. We introduce the Change Blindness (CB) Database with object changes in 130 colored images of natural indoor scenes. The size and eccentricity are provided for all the changes as well as reaction time data from a baseline experiment. In addition, we have two specialized satellite databases that are subsets of the 130 images. In one set, changes are seen in rooms or in mirrors in those rooms (Mirror Change Database). In the other, changes occur in a room or out a window (Window Change Database). Both the sets have controlled background, change size, and eccentricity. The CB Database is intended to provide researchers with a stimulus set of natural scenes with defined stimulus parameters that can be used for a wide range of experiments. The CB Database can be found at http://search.bwh.harvard.edu/new/CBDatabase.html .


Asunto(s)
Bases de Datos Factuales , Estimulación Luminosa/métodos , Percepción Visual , Humanos , Tiempo de Reacción
9.
Psychon Bull Rev ; 22(4): 980-6, 2015 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-25407658

RESUMEN

Is an object reflected in a mirror perceived differently from an object that is seen directly? We asked observers to label "everything" in photographs of real-world scenes. Some scenes contained a mirror in which objects could be seen. Reflected objects received significantly fewer labels than did their nonreflected counterparts. If an object was visible only as a reflection, it was labeled more often than a reflected object that appeared both as a reflection and nonreflected in the room. These unique reflected objects were still not labeled more often than the unique nonreflected objects in the room. In a second experiment, we used a change blindness paradigm in which equivalent object changes occurred in the nonreflected and reflected parts of the scene. Reaction times were longer and accuracy was lower for finding the changes in reflections. These results suggest that reflected information is easily discounted when processing images of natural scenes.


Asunto(s)
Atención/fisiología , Percepción Visual/fisiología , Adulto , Femenino , Humanos , Masculino , Persona de Mediana Edad , Tiempo de Reacción , Adulto Joven
10.
Front Psychol ; 4: 506, 2013.
Artículo en Inglés | MEDLINE | ID: mdl-24009590

RESUMEN

A longstanding goal of computer vision is to build a system that can automatically understand a 3D scene from a single image. This requires extracting semantic concepts and 3D information from 2D images which can depict an enormous variety of environments that comprise our visual world. This paper summarizes our recent efforts toward these goals. First, we describe the richly annotated SUN database which is a collection of annotated images spanning 908 different scene categories with object, attribute, and geometric labels for many scenes. This database allows us to systematically study the space of scenes and to establish a benchmark for scene and object recognition. We augment the categorical SUN database with 102 scene attributes for every image and explore attribute recognition. Finally, we present an integrated system to extract the 3D structure of the scene and objects depicted in an image.

11.
J Exp Psychol Hum Percept Perform ; 38(5): 1315-27, 2012 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-22612057

RESUMEN

Previous research using the contextual cuing paradigm has revealed both quantitative and qualitative differences in learning depending on whether repeated contexts are defined by letter arrays or real-world scenes. To clarify the relative contributions of visual features and semantic information likely to account for such differences, the typical contextual cuing procedure was adapted to use meaningless but nevertheless visually complex images. The data in reaction time and in eye movements show that, like scenes, such repeated contexts can trigger large, stable, and explicit cuing effects, and that those effects result from facilitated attentional guidance. Like simpler stimulus arrays, however, those effects were impaired by a sudden change of a repeating image's color scheme at the end of the learning phase (Experiment 1), or when the repeated images were presented in a different and unique color scheme across each presentation (Experiment 2). In both cases, search was driven by explicit memory. Collectively, these results suggest that semantic information is not required for conscious awareness of context-target covariation, but it plays a primary role in overcoming variability in specific features within familiar displays.


Asunto(s)
Atención/fisiología , Movimientos Oculares/fisiología , Aprendizaje/fisiología , Reconocimiento en Psicología/fisiología , Percepción Visual/fisiología , Adulto , Percepción de Color/fisiología , Señales (Psicología) , Medidas del Movimiento Ocular/instrumentación , Humanos , Pruebas Neuropsicológicas , Reconocimiento Visual de Modelos/fisiología , Tiempo de Reacción/fisiología , Semántica
12.
Front Psychol ; 3: 13, 2012.
Artículo en Inglés | MEDLINE | ID: mdl-22347200

RESUMEN

According to common wisdom in the field of visual perception, top-down selective attention is required in order to bind features into objects. In this view, even simple tasks, such as distinguishing a rotated T from a rotated L, require selective attention since they require feature binding. Selective attention, in turn, is commonly conceived as involving volition, intention, and at least implicitly, awareness. There is something non-intuitive about the notion that we might need so expensive (and possibly human) a resource as conscious awareness in order to perform so basic a function as perception. In fact, we can carry out complex sensorimotor tasks, seemingly in the near absence of awareness or volitional shifts of attention ("zombie behaviors"). More generally, the tight association between attention and awareness, and the presumed role of attention on perception, is problematic. We propose that under normal viewing conditions, the main processes of feature binding and perception proceed largely independently of top-down selective attention. Recent work suggests that there is a significant loss of information in early stages of visual processing, especially in the periphery. In particular, our texture tiling model (TTM) represents images in terms of a fixed set of "texture" statistics computed over local pooling regions that tile the visual input. We argue that this lossy representation produces the perceptual ambiguities that have previously been as ascribed to a lack of feature binding in the absence of selective attention. At the same time, the TTM representation is sufficiently rich to explain performance in such complex tasks as scene gist recognition, pop-out target search, and navigation. A number of phenomena that have previously been explained in terms of voluntary attention can be explained more parsimoniously with the TTM. In this model, peripheral vision introduces a specific kind of information loss, and the information available to an observer varies greatly depending upon shifts of the point of gaze (which usually occur without awareness). The available information, in turn, provides a key determinant of the visual system's capabilities and deficiencies. This scheme dissociates basic perceptual operations, such as feature binding, from both top-down attention and conscious awareness.

13.
Perception ; 40(1): 91-4, 2011.
Artículo en Inglés | MEDLINE | ID: mdl-21513187

RESUMEN

The American painter Gilbert Stuart (1755 - 1828) painted many prominent individuals, including the first six presidents of the United States, and several of his later subjects were also photographed. We compared these photographs to Stuart's portraits of the same individuals and characterized the differences between Stuart's representations and his subjects' actual appearances. This allows us to create more photographic likenesses of other Stuart subjects, including individuals who lived before the invention of photography.


Asunto(s)
Cara/anatomía & histología , Personajes , Retratos como Asunto , Expresión Facial , Historia del Siglo XVIII , Historia del Siglo XIX , Humanos , Pinturas , Fotograbar/historia , Estados Unidos
14.
Vis cogn ; 17(6-7): 945-978, 2009 Aug 01.
Artículo en Inglés | MEDLINE | ID: mdl-20011676

RESUMEN

How predictable are human eye movements during search in real world scenes? We recorded 14 observers' eye movements as they performed a search task (person detection) in 912 outdoor scenes. Observers were highly consistent in the regions fixated during search, even when the target was absent from the scene. These eye movements were used to evaluate computational models of search guidance from three sources: saliency, target features, and scene context. Each of these models independently outperformed a cross-image control in predicting human fixations. Models that combined sources of guidance ultimately predicted 94% of human agreement, with the scene context component providing the most explanatory power. None of the models, however, could reach the precision and fidelity of an attentional map defined by human fixations. This work puts forth a benchmark for computational models of search in real world scenes. Further improvements in modeling should capture mechanisms underlying the selectivity of observer's fixations during search.

15.
Percept Psychophys ; 70(7): 1366-78, 2008 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-18927019

RESUMEN

Because the importance of color in visual tasks such as object identification and scene memory has been debated, we sought to determine whether color is used to guide visual search in contextual cuing with real-world scenes. In Experiment 1, participants searched for targets in repeated scenes that were shown in one of three conditions: natural colors, unnatural colors that remained consistent across repetitions, and unnatural colors that changed on every repetition. We found that the pattern of learning was the same in all three conditions. In Experiment 2, we did a transfer test in which the repeating scenes were shown in consistent colors that suddenly changed on the last block of the experiment. The color change had no effect on search times, relative to a condition in which the colors did not change. In Experiments 3 and 4, we replicated Experiments 1 and 2, using scenes from a color-diagnostic category of scenes, and obtained similar results. We conclude that color is not used to guide visual search in real-world contextual cuing, a finding that constrains the role of color in scene identification and recognition processes.


Asunto(s)
Percepción de Color , Percepción Visual , Señales (Psicología) , Humanos , Tiempo de Reacción , Reconocimiento en Psicología
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...