Visual search

Visual search is a concept that refers to the process of identifying a target object or image among a number of competing distractors. Several factors can complicate the search process, including the number of distractors, how similar the distractors are to the target, how much attention the searcher is applying to the process, and whether there is some context to guide the search process. Visual searches are part of everyday behavior, yet they are a very complex skill that involves many separate components. Computer scientists realized this when they began attempting to develop software capable of replicating the human ability to conduct visual searches.

Background

The first significant efforts to investigate how the visual search process works in the human brain were conducted in the 1970s by English-born psychologist Anne Treisman. Treisman was initially interested in how the ear and brain processed auditory stimulation, and she expanded this to studying how visual stimuli are processed. Treisman determined that different aspects of an object—for instance, its size, color, shape, and movement—are processed in different parts of the brain. She then sought to discover how the brain creates a single image from these different features. She developed the feature integration theory (FIT) to explain how the brain binds the different aspects of an object into one cohesive image.

This study led her to investigate how people distinguish objects from one another. Treisman studied how the brain uses visual cues, past experiences, context, and other factors to locate objects and tell one from another. Her research provided not only greater understanding of how the processes of sight and visual attention function but also helped experts in other fields. For instance, her work helped engineers who design traffic and emergency signs find ways to make the warnings more noticeable.

Overview

Seeing an object goes far beyond the biological function of light entering the eye and stimulating nerves to send signals to the brain. Researchers have determined those signals are sent to different parts of the brain, including areas that determine how far away the object is and decode its color, size, shape, and other features. The brain must then interpret all of this information and bind it together into one cohesive image. This happens even when the current object differs in some way from the person's previous experience with the object. For instance, houses with one or two stories all register as houses; cars are cars whether they are black or blue or sports cars or antique models; a cat id recognizable whether it is standing, sitting, lying down, or half hidden behind a box. This also happens in a fraction of a second as the brain completes this complicated analysis and the object or image is recognized.

Researchers have found that a number of factors influence how the brain processes the complex input necessary. They have determined that the more information the brain has about the target object, the easier it is to identify the object. The brain is set up to look for an object's shape, its orientation, and its size. Other factors are not part of the initial consideration, such as color combinations and factors such as how parts of the object intersect (does it form a T or an X, for example). The brain also uses previous experiences to help in the search process. For instance, while scanning a room looking for a set of keys, the search will focus on the lower half of the room, including the tops of furniture and the floor, but ignore places such as the sides and backs of chairs because experience indicates it is not likely the keys will be found there.

A number of factors have been identified as part of the process to provide guidance that helps many searches take a fraction of a second as opposed to many seconds or even minutes. One is the concept that some areas of possibility are more likely to contain the search target. A person trying to find the way to an appointment will look for directional signs on the wall or suspended from the ceiling because that is where the signs are most likely to be located. Knowing the parameters of the search is very useful. A person looking at pictures of clocks with all the hands in different positions and asked to find the ones that are similar will spend much more time on the task than someone who is told to find the clocks indicating the quarter hour.

The number of distractors that are present can also complicate a search. In the previously mentioned clock scenario, all clocks that are not indicating the quarter hour are distractors, or something that makes the target harder to identify. Nearly anything can be a distractor. For a person waiting for a friend at the airport, all the other people are distractors, for instance.

Researchers have also identified the degree of likelihood that the target will be found to be a factor in visual searches. The difficulty of finding the object is not likely to deter someone who is searching for something like a wedding ring dropped in deep-pile carpeting. However, this has been identified as a challenge for people such as medical professionals who scan hundreds of samples a day looking for tumors or other signs of disease. It is also an issue for workers whose job it is to scan for contraband objects such as weapons and bombs. The mere fact that they might review thousands of images without seeing a target image makes it more likely that they will overlook the target when it appears. This is partially a function of attention and partially a function of what is known as the prevalence effect. This is the tendency of the brain to not see something because it does not expect to see it; hundreds of visual searches that end without finding a target essentially condition the brain to not see the target.

For the most part, visual search processes happen subconsciously, without the searcher being aware of the factors that are part of the search. Researchers are working to better understand how these processes occur so that people whose job it is to successfully complete visual searches can be trained to do so more effectively. They also seek to understand how the human brain completes these processes in an effort to program computers to assist with such tasks as finding tumors, identifying terror threats, and seeking other potential dangers.

Bibliography

"Anne Marie Treisman." Princeton University, dof.princeton.edu/about/clerk-faculty/emeritus/anne-marie-treisman. Accessed 30 Nov. 2017.

Conkle, Ann. "Visual Search Gets Real." Association for Psychological Science, www.psychologicalscience.org/observer/visual-search-gets-real. Accessed 30 Nov. 2017.

Davis, E.T., and J. Palmer. "Visual Search and Attention: An Overview." Spatial Vision, vol. 17, no. 4–5, 2004, pp. 249–55.

Eckstein, Miguel P. "Visual Search: A Retrospective." Journal of Vision, Dec. 2011, jov.arvojournals.org/article.aspx?articleid=2191835. Accessed 30 Nov. 2017.

Kristjánsson, Árni. "Reconsidering Visual Search." i-Perception, vol. 6, no. 6, 2015, journals.sagepub.com/doi/pdf/10.1177/2041669515614670. Accessed 30 Nov. 2017.

"Learning in Visual Search." American Psychological Association, 14 Aug. 2014, www.apa.org/pubs/highlights/peeps/issue-27.aspx. Accessed 30 Nov. 2017.

"Visual Search." Cognitive Science Software, Indiana University Bloomington, cognitrn.psych.indiana.edu/cogscisoftware/visualsearch/index.html. Accessed 30 Nov. 2017.

"Visual Search for Features and Conjunctions." Jove, www.jove.com/science-education/10062/visual-search-for-features-and-conjunctions. Accessed 30 Nov. 2017.