How It Works

Photography is the process of capturing light onto a photosensitive medium — whether that's a silicon sensor or a strip of silver-halide film — and the physics governing that process haven't changed since the first daguerreotypes. What has changed is the complexity layered on top of that physics. This page breaks down the core mechanism of how a camera captures an image, the sequence from shutter press to finished file, the roles different components play, and the variables that determine whether the result is what the photographer intended.

The basic mechanism

At its most fundamental level, a camera is a light-tight box with a controlled opening. When the shutter opens, photons pass through the lens, which bends them into a focused pattern, and strike the imaging surface at the back. That surface — film or sensor — records the intensity and color of the light at each point across its area.

On a digital sensor, millions of photosites (commonly called pixels) each collect photons and convert them into electrical charge. A typical full-frame sensor measures 36mm × 24mm and might contain 24 to 61 million of those photosites, depending on the camera model. Each photosite is covered by a colored filter — red, green, or blue — arranged in a pattern called a Bayer array, named after Kodak engineer Bryce Bayer who patented it in 1976. The camera's processor then demosaics that raw data, interpolating full-color information for every pixel.

Film works differently: silver halide crystals suspended in an emulsion react chemically to light, forming a latent image that chemical development makes visible. The comparison between digital and film photography reveals just how divergent these two paths become after that initial moment of exposure — same physics, entirely different downstream processes.

Sequence and flow

The sequence from pressing the shutter to having a usable image involves more steps than most people expect.

  1. Metering — The camera (or photographer) evaluates scene brightness, often using evaluative, center-weighted, or spot metering modes, to calculate a recommended exposure.
  2. Autofocus — Phase-detection or contrast-detection systems find the focal plane. Phase-detection, standard on most interchangeable-lens cameras, measures the convergence of two light beams split by a mirror or on-sensor microprisms, resolving focus in milliseconds.
  3. Shutter actuation — The mechanical or electronic shutter opens for the set duration, from as long as 30 seconds to as brief as 1/8000 of a second on many modern bodies.
  4. Sensor readout — Electrical charge from each photosite is read row by row and converted to digital values by the analog-to-digital converter (ADC).
  5. Processing — The camera's DIGIC, EXPEED, or BIONZ processor (depending on Canon, Nikon, or Sony) applies demosaicing, noise reduction, sharpening, and color rendering.
  6. Write to storage — The resulting data is written as a RAW file, a JPEG, or both, to the memory card. The RAW vs JPEG distinction at this stage has significant downstream consequences for editing latitude.

Roles and responsibilities

The lens, sensor, and processor are the three principals — and they divide labor in ways that matter.

The lens is responsible for optical quality: resolving fine detail, controlling aberrations, and determining the field of view. A 50mm f/1.8 prime and a 50mm f/1.8 zoom may share the same focal length and maximum aperture, but their optical formulas differ substantially, typically giving the prime an edge in sharpness and rendering. The lens types and uses guide covers these distinctions in depth.

The sensor determines dynamic range (how much tonal information it can capture between pure black and clipped white), ISO performance (how well it handles amplified signal at high sensitivities), and resolving power. These are hardware constraints no amount of in-camera processing can fully overcome.

The processor handles everything computational — and on modern mirrorless bodies, this includes real-time subject tracking, computational noise reduction, and face/eye detection running at 30 frames per second or more. The photographer sits above all three, directing the exposure triangle — aperture, shutter speed, and ISO — which is the control layer that shapes what the hardware actually captures. The exposure triangle explains how those three variables interact and trade off against each other.

What drives the outcome

Two cameras with identical specifications can produce meaningfully different results depending on three factors: light, intent, and post-processing.

Light is the raw material. Direction, quality (hard vs. soft), color temperature, and intensity determine what the sensor has to work with before any setting is applied. Lighting in photography treats this as its own discipline — because it is.

Intent refers to the compositional and technical decisions made before the shutter fires: focal length choice, depth of field, moment of capture. A 400mm telephoto compresses background and subject together; a 16mm wide-angle separates them dramatically. Neither is correct — both are tools for a specific visual argument. The photography composition rules page anchors these decisions in established visual principles.

Post-processing is where the captured data is interpreted. A RAW file straight from the sensor looks flat and slightly greenish because no rendering has been applied. Color grading, tonal adjustments, and sharpening in tools like Lightroom or Capture One translate that raw data into the photographer's actual vision. The gap between what the sensor captured and what the eye saw is closed here.

Understanding this chain — from photon to pixel to processed file — is the foundation that makes everything else in photography legible. The full scope of what photography encompasses extends far beyond the mechanical moment of capture, but that moment is where everything begins.