Resources
One of the trends that can be seen in microscopy in recent years has been the interest in imaging sensors with larger fields of view and smaller pixel sizes to maintain image detail when working at lower magnifications such as 20x. But are smaller pixels always better, or are there still conditions in which larger pixels are advantageous? In this article we consider pixel size and whether larger pixel sizes are still useful, and if so, where may this apply.
The role of a pixel in a sensor is to collect photons within a subunit of the imaging area and subsequently convert these to an electrical signal. By digitizing these signals, an image can be recreated from the values received at each pixel. The more efficient at gathering light and converting this light to an electrical signal, the more sensitive the detector can be. This has been a key driver of CCD and CMOS sensor technologies with processes such as back-illumination and use of microlens being examples of ways to boost the efficiency of these devices. The latest sensors available have quantum efficiencies reaching up to 95% i.e. 95% of incoming photons are converted into electrons in the photosensitive region of the silicon compared to devices a little over 10 years ago being limited to around 60% peak QE.
Figure 1: Modern sensor formats allow for high sensitivity by a combination of high photon collection efficiency and conversion and low noise electronics. This illustration shows a back-illuminated sensor in which the circuitry is beneath the photosensitive region and thus does not block light from reaching the region in which photons may be converted to electrons.
Developments in sensor design and fabrication has enabled ever more smaller pixels to be packed into sensors. While more pixels can be better, there are trade-offs so the pixel size must be balanced to suit the needs of the application. There are three main aspects of how pixel size may impact imaging in addition to advances in other sensor technology enhancements such as back-illumination. These are summarized in table 1:
Sensor Parameter | Image Parameter | Description |
Photon Collection Efficiency | Signal to noise | The more sensitive a sensor is, the more low level information can be obtained. This can be the difference between detecting a signal and not. A sensor with a high photon collection ability will help collect more signal which will boost the signal against the noise of the sensor itself. |
Image sampling “Resolution” | Image Detail | Resolution is the ability to distinguish two objects from their Airy pattern. The resolution is strictly defined by what is possible through the microscope via the Rayleigh equation: 0.61 x Wavelength/objective NA. The imaging detector should have sufficient sampling of the image to meet Nyquist criteria and maintain that resolution and thus the image detail. |
Well depth | Dynamic Range | Dynamic range is the range between the lowest and highest levels of signal within one image. Dynamic range of the sensor can be derived from the well depth of the sensor divided by the noise floor. This assumes that the sensor supporting electronic design is optimised to handle the potential dynamic range and do so in a linear relationship i.e. bit-depth and amplifier designs |
Table 1: Sensor parameters linked to pixel size and relationship to the image
Matching pixel sizes to microscope field of view and maintaining resolution is discussed more comprehensively in the article: Optimizing field of View and Resolution for Microscopy. In the following sections we will focus on pixel size and photon collection efficiency.
In table 2 we can see the pixel sizes of a range of current commercially available high performance imaging sensors that may be used for microscopy:
Sensor Technology | Camera | Pixel size (µm) | Pixel Area (µm2) | Difference in pixel photon collection area |
CMOS | Back-illuminated CMOS | 4.6 | 21.2 | 1x |
Front illuminated sCMOS e.g. Zyla 4.2PLUS | 6.5 | 42.25 | 2x | |
Back-illuminated sCMOS e.g. Sona-6 |
6.5 | 42.25 | 2x | |
Back-illuminated sCMOS e.g. Sona-11 | 11 | 121 | 5.7x | |
EMCCD | Back-illuminated EMCCD e.g. iXon 888 | 13 | 169 | 8x |
Back-illuminated EMCCD e.g. iXon 897 | 16 | 256 | 12x | |
CCD | Back-illuminated CCD e.g. iKon-M | 13 | 169 | 8x |
Back-illuminated CCD e.g. iKon-L | 13.5 | 182.25 | 8.6 |
Table 2: The photon collection areas for each pixel for a range of different camera models
In general, CMOS sensors have smaller pixel sizes than EMCCD and CCD sensors. The CMOS sensor architecture also allows higher speeds and larger sensor sizes than CCD based designs. These qualities suit general fluorescence imaging applications allowing good image detail at common magnifications for cell biology studies and have made sCMOS the dominant detector technology for many microscopists.
Despite the rise of sCMOS detectors, EMCCD and CCD detectors have remained as the best option for some of the most challenging imaging applications and this looks set to continue for some time to come:
• EMCCD cameras use Electron Multiplication to effectively eliminate readout noise. This unique feature makes this sensor technology ideal for making precise quantitative measurements down to even a single photon such as single molecule imaging, but also for live-cell confocal imaging and super resolution systems.
• Deep cooled CCD cameras feature orders of magnitude less dark current than CMOS designs making them perfectly suited for long exposure luminescence experiments that are dark current limited and the slower readout of CCD is not a factor.
The large pixel sizes that these EMCCD and CCD cameras have also plays an important factor in why they remain suitable for these imaging scenarios. When photon collection is the priority, a larger pixel acts effectively as a larger catchment area for photons- feeding in more signal to the sensor. This is why the latest EMCCD sensor designs feature pixel sizes in the range of 13-16 µm which are natively suited to photon collection at high magnification e.g. 100x. Note that EMCCD cameras with smaller pixel sizes are available, but they have not gained any market traction as they do not provide application benefits over large pixel EMCCD, or to sCMOS camera options. In the following illustration, an iXon Ultra 888 EMCCD camera has a 4x higher native pixel area than a typical sCMOS camera and an 8x larger pixel area than the CMOS sensor.
Figure 2: Large pixel sizes are used in EMCCD and CCD cameras to gain advantage of the superior photon collection efficiency and improved probability of photon being detected. This example illustrates the relative sizes of 13 µm, 6.5 and 4.6 µm pixel.
There are some back-illuminated sCMOS cameras available with a pixel size of 11 µm pixel - which is a relatively large pixel when compared to typical CMOS cameras. These camera models, such as the Sona-11 sCMOS series, can take advantage of the light gathering power 3-fold higher than the 6.5 µm pixel size typically found in sCMOS cameras.
For some applications such as single molecule detection, and for high magnification, the larger pixel sCMOS cameras like the Sona-11 can offer some benefits:
However, for the more challenging single molecule experiments, the larger pixel size and high QE is not always enough. Electron multiplication process unique to EMCCD sensors boost the signal many fold prior to readout so the signal is far above the noise floor of the camera readout electronics and in practice proven to operate at photon levels below what is possible from any CMOS technology
Here we compare 2 cameras with all factors the same (QE, read noise, dark current) other than their pixel size and compare the effect of pixel size on relative the signal to noise performance as light of a given illumination falls on each sensor.
Camera | Pixel Size (µm) | Pixel Area (µm) | Signal (photons) | Noise |
1 | 6.5 | 42.25 | 10 | 2 |
2 | 13 | 169 | 40 | 2 |
Table 3: Simplified Signal to Noise comparison to illustrate the increase in the signal component provided by pixel size.
In the simplest comparison, camera 2 would look to provide a significantly higher signal to noise ratio since the signal component has increased for each pixel. To consider signal to noise for imaging cameras we need to use the following signal to noise equation to model the different signal and noise components involved:
Using this information we can plot the signal to noise against the number of photons, accounting for illumination intensity over the sensor by expressing photons on an area basis – in this example we will express per 13 µm2 sensor area:
Figure 2: A comparison the effect of pixel size on signal to noise performance of 2 sensors with all factors the same other than pixel size with incident light as photons normalised over per 13 µm.
From this we can see that the larger pixel does indeed allow for a higher signal to noise - some ~2.5x higher at 10 photons per 13µm2. This will translate to the following benefits:
We can then apply this to real-world camera examples:
Figure 3: Cameras developed for sensitivity and highest possible signal to noise feature large pixel sizes. Illumination corrected for pixel size by normalizing per 13 µm area.
In the comparison illustrated in figure 3 we can see that the cameras with larger pixels provide the better signal to noise figures such as the iXon Ultra 888 and Sona-11 models which aligns with their intended applications, that prioritize sensitivity and detection, over spatial resolution at low light levels. The other CMOS models with smaller pixels do not provide as high signal to noise ratios even though they have low noise levels. They will provide improved spatial resolution.
Smaller pixels can be binned – whereby the signal in a number of pixels may be pooled together to increase the overall signal level. A common example would be 2x2 binning which would group the signal for 4 pixels together as shown in figure 3. This works effectively for CCD in boosting speed and sensitivity as the serial nature of CCD allows each pixel value to be added before a read noise is added. Binning also works especially well outside of imaging for spectroscopy measurements as vertical binning may be applied without concern of losing resolution in the y axis.
Figure 4: 2x2 binning applied to sensor to boost photon collection across 4 pixels and the signal to noise ratio at the expense of spatial resolution.
The architecture of CMOS however is different. Each row has to be read out sequentially on a row-by-row basis before binning - meaning that for the CMOS sensors in use today binning will not deliver the benefits to signal to noise (and speed) that it does for CCD. Digital binning for sCMOS still combines he values for 4 pixels in the case of 2x2 binning, but the read noise also increases 2-fold since the 2 pixel rows are read out.
Figure 5: While binning does not work as well for CMOS as CCD, it still allows a boost in signal to noise compared to without at the expense of spatial resolution.
For the smaller 4.6 µm pixels of the CMOS sensor 3x3 binning would be required to obtain the same photon collection area as a 13 µm pixel. 3x3 binning would thus increase read noise in this sensor type by a factor of 3.
Ideally the sensor size should fit within the uniform region available to the microscope and the pixel size would match the objective and magnification used for imaging. However, this is sometimes not the case for different combination of sensors, microscopes and application requirements. Additional post objective magnification may be used to avoid vignetting effects, or expand the field of view to cover that of a larger sensor area. The other application of additional magnification is to help meet optimal sampling of the image to meet or exceed Nyquist criteria e.g. adding a 2x lens with reduce the pixel size by x2 and help improve spatial resolution, but reduce the field of view. Thus, while useful, there are a number of considerations of using additional lenses in the optical system. This topic is discussed further in the article: Optimizing field of View and Resolution for Microscopy.
The final aspect of imaging performance that is linked to pixel size that can be beneficial is the well depth of the sensor and how this can be used to allow for high dynamic ranges. When implemented correctly, larger pixels often allow larger signal handing capacities than is possible using smaller pixels.
Sensor Technology | Camera | Pixel size (µm) | Well Depth (e-) |
CMOS | Back-illuminated CMOS | 4.6 | 7,000 |
Front illuminated sCMOS Zyla 4.2PLUS | 6.5 | 30,000 | |
Back-illuminated sCMOS e.g. Sona-6 | 6.5 | 55,000 | |
Back-illuminated sCMOS e.g. Sona-11 | 11 | 85,000 | |
EMCCD | Back-illuminated EMCCD e.g. iXon 888 | 13 | 80,000 |
Back-illuminated EMCCD e.g. iXon 897 | 16 | 180,000 | |
CCD | Back-illuminated CCD e.g. iKon-M | 13 | 100,000 |
Back-illuminated CCD e.g. iKon -L | 13.5 | 150,000 |
Table 4: A comparison of the well depth of a selection of cameras. Note that maximum dynamic range may be restricted to certain modes available for each of the different camera models so is not presented here.
The impact in imaging terms will be that using small pixels may mean a low dynamic range that can result in oversaturation and lose of image information. Therefore, for applications in which a high dynamic range is important a well optimised sensor with larger pixel sizes may be more suitable than one with smaller pixels.
Imaging scenarios one may require high dynamic range include:
A further benefit of wide dynamic range is from a practical sense – a sensor with a wide dynamic range is more often in range and avoids saturation requiring less adjustment of settings such as exposure.
While there is interest in imaging cameras with small pixels to preserve image detail at low magnifications there is still very much the need for cameras with larger pixel sizes for low light applications.
There is no one camera that does all imaging scenarios better than every other camera. Thankfully there are now a range of cameras that can be selected to suit different applications that need for example the highest spatial or temporal resolution, or the best possible sensitivity. This means that it is important to consider the application needs first and foremost and not assume that the latest camera model with the largest sensor and most megapixels will be the best option.