Convolutional Neural Networks | How Does AI Learn to See Like a Human?

What is a Convolutional Neural Network (CNN)?

How do CNNs and the visual cortex process information hierarchically?

A Convolutional Neural Network (CNN) is a type of artificial neural network inspired by the organization of the animal visual cortex. Its structure is based on a concept called hierarchical processing. In the human brain, visual information is processed in stages. The primary visual cortex (V1) first detects very simple features, such as edges, lines of specific orientations, and colors. As this information moves to subsequent areas (V2, V4, IT cortex), these simple features are combined to form more complex representations, like shapes, textures, and eventually whole objects such as faces or cars. CNNs replicate this exact process. The initial layers of the network learn to detect basic edges and color gradients from raw pixel data. The output of these layers, called feature maps, are then fed into deeper layers, which learn to combine the simpler features into more intricate patterns like corners, circles, or textures. This hierarchical assembly continues until the final layers can identify complex objects, demonstrating a clear parallel with the brain's own visual processing pathway. This layered approach allows CNNs to build a rich, abstract understanding of an image from simple, fundamental components.
notion image

What are filters in CNNs and how do they relate to neurons?

In a CNN, "filters" (also known as kernels) are the primary tools for feature detection. A filter is a small matrix of numbers that slides over the input image to produce a feature map. Each filter is designed to detect a specific pattern. For example, one filter might be tuned to detect horizontal edges, while another detects vertical edges or a specific color pattern. This is directly analogous to the concept of receptive fields in the biological visual cortex. In the brain, a single neuron in the visual cortex only responds to light stimuli in a limited region of the visual field. Furthermore, these neurons are often tuned to specific features, such as lines of a particular orientation. Just as a specific neuron fires only when it "sees" the pattern it's tuned for, a CNN filter produces a high activation value in the feature map when it slides over a region of the image that matches its pattern. In essence, the filters in the early layers of a CNN function as specialized simple-cell neurons.

Exploring the Functional Parallels

Is the 'convolution' operation in CNNs biologically plausible?

The core mathematical operation in a CNN is the "convolution." This operation involves sliding a filter across the entire image and computing the dot product at each location. The result is a feature map that highlights where the specific feature detected by the filter appears. This process is considered biologically plausible because it mirrors how neurons in the visual cortex have receptive fields that overlap and cover the entire visual field. Each neuron is tuned to a feature, and the brain builds a complete map of these features. The convolution's shared weights—where the same filter is used across the entire image—is a key aspect. This is analogous to how the brain assumes that a feature, like a vertical edge, is equally important regardless of where it appears in the visual field. This property, known as translation invariance, is fundamental to both CNNs and human vision.
notion image

Does the brain have a 'pooling' mechanism like CNNs?

In CNNs, a "pooling" (or subsampling) layer often follows a convolutional layer. Its function is to reduce the spatial dimensions of the feature map. For example, a max-pooling layer takes a small window of the feature map and outputs only the maximum value. This makes the network's representation of the feature more robust to its exact location; as long as the feature is detected somewhere within the pooling window, its presence is recorded. This creates a degree of spatial invariance. In the brain, complex cells in the visual cortex exhibit a similar property. Unlike simple cells that respond to a line in a very specific location, complex cells respond to a line of a specific orientation anywhere within their larger receptive field. This ability to abstract the presence of a feature away from its precise position is a key function of pooling in CNNs and is considered a direct parallel to the function of complex cells in the brain, helping us recognize objects even if they move slightly.

Limitations and Key Differences

How different are CNNs from the actual visual cortex?

While the inspiration is clear, significant differences exist between CNNs and the brain's visual cortex. Firstly, the brain is vastly more energy-efficient and can learn from significantly less data. A human can often recognize an object class after seeing just one or two examples (one-shot learning), whereas a CNN requires thousands or millions of labeled images for training. Secondly, the brain's learning mechanisms are not fully understood, but they appear to be more complex than the backpropagation algorithm used by nearly all CNNs. Backpropagation requires a global error signal to be sent backward through the network, a process for which there is limited direct biological evidence. Furthermore, vision in the brain is not a purely feedforward process like in a standard CNN. The brain has extensive feedback connections, allowing higher-level cognitive areas to influence and guide processing in lower-level sensory areas. Finally, biological vision is deeply integrated with other senses, memory, and context, providing a much richer and more robust understanding of the world than the isolated, task-specific function of a typical CNN.
notion image