Content from Opening and Checking an Image


Last updated on 2025-03-21 | Edit this page

Overview

Questions

  • How do I open an image file in Python for processing?
  • How can I explore an image once it’s open?

Objectives

  • Open an image with skimage
  • Discuss how to open proprietary image formats
  • Display an opened image to the screen

Opening an image


At its core, an image is a multidimensional array of numbers, and as such can be opened by programs and libraries designed for working with this kind of data. For Python, one such library is scikit-image. This library provides a function called imread that we can use to load an image into memory. In a new notebook cell:

PYTHON

from skimage.io import imread
image = imread('data/FluorescentCells_3channel.tif')

If you saved your images to a different location, you will need to change the file path provided to imread accordingly. Paths will be relative to the location of your .ipynb notebook file.

To view things in Python, usually we use print(). However if we try to print this image to the Jupyter console, instead of the image we get something that may be unexpected:

OUTPUT

> print(image)
array([[[ 16,  50,   0],
        [ 15,  44,   0],
        [ 18,  40,   1],
        ...,
        [  1,  15,   2],
        [  1,  15,   2],
        [  1,  15,   2]]], dtype=uint8)

Python has loaded and stored the image as a Numpy array object of numbers, and print() displays the string representation or textual form of the data passed to it, which is why we get a matrix of numbers printed to the screen.

If we want to see what the image looks like, we need tell Python to display it as an image. We can do this with the imshow function from matplotlib, which you may already be familiar with as a library for drawing plots and graphs, but it can also display images.

PYTHON

import matplotlib.pyplot as plt

plt.set_cmap('gray')  # by default, single-channel images will now be displayed in greyscale
plt.imshow(image)

You should now see the image displayed below the current cell:

Displaying an image with imshow

Since images are multi-dimensional arrays of numbers, we can apply statistical functions to them and extract some basic metrics. Numpy arrays have methods for several of these already, including the image’s shape, data type and minimum/maximum values:

PYTHON

print(image.shape, image.dtype, image.min(), image.max())
# returns: (512, 512, 3) uint8 0 255

This shows that the image has a data type of uint8, it contains values between 0 and 255 and that it is in three dimensions. We can reasonably infer that the two 512 numbers are the X and Y axes. The third axis in most cases will represent a number of channels.

We can select a single channel by indexing the array:

PYTHON

plt.imshow(image[:, :, 2])

Here, we select the entire X and Y axes using : with no numbers around them, and the last channel (remember that Python counts from 0).

Channels, series and stacks


Images can consist of more than two axes. The first two axes are usually X and Y, but if there is a third axis, then this could be one of several things:

  • Channel - the image shows different features in the same 2D space. One common example is cell images with different staining for nuclei and membranes, expressed as different colours.
  • Time series - the image is a collection of 2D frames taken at different points in time.
  • Z-stack - essentially a series of 2D images piled up on top of each other in 3D space.

It’s usually easy enough to tell that you’re looking at a colour channel from looking at the image directly, but it may be more difficult to to distinguish a Z or a timepoint axis from the data alone. If you don’t know exactly how the images were generated, it’s a good idea to consult documentation or metadata.

Exercise 1: Loading an image

Load the test image ‘FluorescentCells_3channel.tif’:

  • Try the same as the example above, but display one of the other channels
  • Save your single channel to a variable. What happens if you run imshow on channel.T?
  • How can we select part of the image, i.e. crop it? Remember that to do this, we need to select a subset of the X and/or Y axes.

Other channels can be loaded with image[:, :, x], where image is the variable the image is saved to and x is the index of a channel to retrieve.

Next we can use .T to return a transposed version of the image. Running imshow() on this results in an image that is flipped 90°:

PYTHON

channel = image[:, :, 1]
plt.imshow(channel.T)
Transposed image

Remember that Numpy arrays can be sliced and indexed the same way as lists, strings and tuples. Up to this point we’ve been using : to select an entire axis, but we can give it start and end bounds to select part of the X and Y axes, like:

PYTHON

image[:256, 128:384, 1]
2-dimensional slice

Numpy arrays have many more methods available for checking them. Here are just a few to start with:

Image size

image.shape

Object size

image.nbytes

Note that some of these are functions and need to be called with brackets(()), whereas others are simply attributes that do not.

Exercise 2: Memory check

How much memory did it take to load FluorescentCells_3channel.tif?

Running image.nbytes shows that it takes up 786432 bytes, or ~786 kilobytes, or ~0.7 megabytes.

Displaying one channel at a time


We’ve seen from exercise 1 that we can view single channels by indexing the array. We can also show all channels together using a matplotlib figure:

PYTHON

import matplotlib.pyplot as plt
plt.figure(figsize=(12, 6))  # figure size, in inches
nchannels = 3

for i in range(nchannels):
    # Use subplot() to create a multi-image figure with 1 row and 3 columns. We need to increment i by 1
    # because range() counts from 0 but subplot() assumes you're counting from 1.
    plt.subplot(1, nchannels, i+1)
    plt.imshow(image[:, :, i])
    plt.title('Channel ' + str(i))  # add a plot title
    plt.axis(False)              # we just want to show the image, so turn off the axis labels

plt.show()
All image channels

Histograms


Another useful metric in image analysis is an image’s histogram. This can be plotted by flattening the image and passing it to matplotlib:

PYTHON

plt.hist(image[:, :, 0].flatten(), bins=256)

First, we need to select a single channel - since different channels may represent different cell organelles or points in time, we need to ensure that we are comparing like with like. We also need flatten() because we don’t care about the arrangement of the pixels, we just want to sort their values values into bins. Finally, we can use bins to control how many bins the data is split into.

Exercise 3: Histograms

Combine the usage of matplotlib.pyplot.hist() and matplotlib.pyplot.figure() introduced above above and plot a histogram of each of the three channels in FluorescentCells_3channel.tif.

Starting with displaying a single histogram for one channel:

PYTHON

channel_idx = 0
channel = image[:, :, channel_idx]
plt.hist(channel.flatten(), bins=255)
plt.title('Channel ' + str(channel_idx))
plt.show()

We could call this three times, each with a different value for channel_idx, or we can use a for loop:

PYTHON

plt.figure(figsize=(12, 6))  # figure size, in inches
nchannels = image.shape[-1]

for i in range(nchannels):
    channel = image[:, :, i]

    plt.subplot(1, nchannels, i+1)
    plt.hist(channel.flatten(), bins=255)
    plt.title('Channel ' + str(i))  # add a plot title

plt.show()
All image channels

Proprietary formats


Images can come in many formats, but for many of the common ones such as TIFF, PNG and BMP, skimage is smart enough to know how to read each one.

Some image formats are associated with specific instruments or equipment and require specialised packages to open. Depending on your system, these may already be available via import the same as any other Python package. If not, then these should be installed into whatever Python instance you are using.

If using JupyterHub or JupyterLab, go to ‘New’ -> ‘Terminal’. This will open a shell session in a new browser tab, where you can run pip install commands.

Carl Zeiss .czi

Images in .czi format can be opened with the czifile library. In the Terminal you opened above:

SH

$ pip install czifile

Nikon .nd2

SH

$ pip install nd2

Imaris .ims

SH

$ pip install imaris-ims-file-reader

Leica .lif

SH

$ pip install readlif

Exercise 4: Proprietary formats

Load the nd2 package and use it to read the test file ‘Ersi_organoid_WT2.nd2’. Install it if you need to.

What axes are present in the image? Which look most likely to be the X and Y axes? Use imshow to display a single frame from the image.

If not already present, the nd2 package can be installed in the Terminal with pip install nd2. According to its linked documentation, it has an imread function that works in a similar way to the one in skimage, and returns a numpy multidimensional array that we can work with the same way we have before:

PYTHON

import nd2
image = nd2.imread('data/Ersi_organoid_WT2.nd2')
image.shape
# returns: (27, 3, 512, 512)

img.shape shows that there are four axes, (27, 3, 512, 512). The latter two numbers look like the X and Y axes, while the second number looks like a number of colour channels. The first number looks like either a time series or a Z axis. We can show a single frame with:

PYTHON

plt.imshow(image[0, 0, :, :])
ND2 image

As discussed above in exercise 4, it may be difficult to distinguish a time series from a Z axis. You may also notice that here the X/Y axes are the latter two numbers, but in other examples above, the X/Y axes were the first two. This demonstrates the diversity and general lack of consistency in image formatting, and how it’s usually a good idea to find out as much as you can about the image from documentation and metadata before processing it.

Altering the lookup table


Now that we’ve been able to open an image, we can start to explore it and display it in different ways.

The imshow() function can take extra arguments in addition to the image to display. One of these is called cmap, which can apply alternate lookup tables (a.k.a. colour maps):

PYTHON

plt.imshow(image[0, 0, :, :], cmap='viridis')

Skimage uses lookup tables from the plotting library matplotlib. A list of available tables can be obtained with:

PYTHON

from matplotlib import colormaps
print(sorted(colormaps))

Exercise 5: Lookup tables

Go back to the FluorescentCells_3channel.tif image. Display each of its three channels side by side in a matplotlib figure, each in a different colour using cmap=. Use the values in matplotlib.colormaps to select a lookup table for each one.

There will be many ways to do this (and many colour maps to choose from!), but here is one possible solution:

PYTHON

# you'll need to run this again if you overwrote your `image` variable
image = imread('data/FluorescentCells_3channel.tif')
plt.figure(figsize=(12, 6))

plt.subplot(1, 3, 1)
plt.imshow(image[:, :, 0], cmap='Blues')

plt.subplot(1, 3, 2)
plt.imshow(image[:, :, 1], cmap='Oranges')

plt.subplot(1, 3, 3)
plt.imshow(image[:, :, 2], cmap='YlOrBr')
ND2 image

Exercise 5: Lookup tables (continued)

Load hela-cells_rgb.tif and try displaying it with different lookup tables. What results do you get? Why might this be the case?

The resulting image will not look as expected, and in Jupyter the image will appear unchanged. In this case, since the lookup table is being ignored, this would imply that the pixel values do not represent light intensities but rather are explicitly encoded colour values - i.e. it’s an RGB image.

Other notes


Rearranging channels

You may find yourself in a situation where the arrangement of dimensions in your image is incorrect for the processing you wish to perform on it - maybe a function requires that an image be oriented in a particular way. This is where it’s useful to be able to rearrange the dimensions of an image. To do this, we can use Numpy’s moveaxis function:

PYTHON

import numpy
image = imread('data/FluorescentCells_3channel.tif')
print(image.shape)
# returns: (800, 800, 3)
rearranged_image = numpy.moveaxis(image, -1, 0)
print(rearranged_image.shape)
# returns: (3, 800, 800)
print(image.shape)
# returns: (800, 800, 3)

We can see that calling moveaxis on an array gives us a rearranged version of the array given to it - the channel axis that was at the end is now at the front. However, we can see that the original value of the image is unchanged. This is because by default, moveaxis returns a rearranged copy of the image.

The arguments supplied are:

  • The image or array
  • The current position of the dimension to move
  • The position to move that dimension to

In this case, we are moving the dimension at position -1 (i.e. the one at the end) to position 0 (the start).

Pixel size

Pixels are an approximation of an object - knowing that something is 50 pixels long and 50 pixels wide doesn’t tell us anything about its actual size. To make any real-world measurements on the image, we need the image’s pixel size.

To get this, it is necessary to read the image’s metadata. For this we need a different library, imageio. There are a couple of different places we can look:

PYTHON

import imageio
meta = imageio.v3.immeta('data/FluorescentCells_3channel.tif')
props = imageio.v3.improps('data/FluorescentCells_3channel.tif')
print('meta:', meta)
print('props:', props)

immeta() gives us a dict summarising the image, and improps gives an object with the property ‘spacing’. However, there is no guarantee that wither of these will contain any information on pixel size, and the BioimageBook notes that these numbers can be misleading and require interpretation and cross-checking.

Even if immeta() does return a key called ‘unit’, the value may be returned as escaped Unicode:

SH

'unit': '\\u00B5m'

This can be un-escaped with:

SH

>>> m['unit'].encode().decode('unicode-escape')
μm

Here, this would indicate that the image is to be mesaured in micrometres. Combining this with other information that may be available, e.g. improps().spacing, will help you figure out the pixel size.

Key Points

  • Common image formats can usually all be loaded in the same way with skimage
  • Specialised proprietary formats may require specialised libraries
  • Basic metrics of an image include histogram, shape and max/min pixel values
  • These metrics can help tell us how the miage should be analysed
  • Lookup tables can change how a single-channel image is rendered
  • An RGB image contains 3 channels for red, green and blue

Content from Applying Filters


Last updated on 2025-03-21 | Edit this page

Overview

Questions

  • How can I make it possible to isolate features in my images?

Objectives

  • Apply mean and Gaussian filters to an image
  • Select one of these filtered images to use for processing in subsequent chapters

Filters


Images often need to have noise removed in order for the results of further processing to be meaningful. There are many smoothing filters available, each with their own advantages depending on the situation and the image in question. Here are two to get started:

A mean filter will, for each pixel:

  • Create a kernel of pixels around it, in a size/shape of the user’s choosing
  • Take the mean of all pixels within this kernel
  • Assign this new value to the pixel

The kernel in this case can be considered analogous to a 2-dimensional sliding window.

A Gaussian filter is similar to a mean filter, except that pixels near the centre of the kernel will have a greater effect on the result.

You can create a kernel with skimage:

PYTHON

from skimage.morphology import square
kernel = square(3)  # create a 3x3 square kernel

There are other shapes of kernel that can be used, and are documented in the morphology section of the skimage docs.

Note that as of skimage 0.25.0, the square function has been deprecated in favour of a new function, footprint_rectangle:

PYTHON

from skimage.morphology import footprint_rectangle
kernel = square((3, 3))  # also a 3x3 square kernel

Exercise 6: Applying filters

Look at the documentation pages for the mean and Gaussian filters above. Load a frame from the second channel of the test image skimage.data.cells3d:

PYTHON

from skimage.data import cells3d

image = cells3d()[30, 1, :, :]

Build a figure displaying this image in each of the following forms:

  • original image
  • mean filter using a 3x3 square kernel
  • mean filter using a 9x9 square kernel
  • Gaussian filter of sigma = 1
  • Gaussian filter of sigma = 5

How do the different methods compare?

PYTHON

from skimage.data import cells3d
from skimage.filters import gaussian
from skimage.filters.rank import mean
from skimage.morphology.footprints import square

image = cells3d()[30, 1, :, :]
plt.figure(figsize=(10, 12))

plt.subplot(3, 2, 1)
plt.imshow(image)
plt.title('Original')

plt.subplot(3, 2, 2)
plt.imshow(mean(image, footprint=square(3)))
plt.title('3x3 mean filter')

plt.subplot(3, 2, 3)
plt.imshow(mean(image, footprint=square(9)))
plt.title('9x9 mean filter')

plt.subplot(3, 2, 4)
plt.imshow(gaussian(image, sigma=1))
plt.title('Gaussian blur, σ=1')

plt.subplot(3, 2, 5)
plt.imshow(gaussian(image, sigma=5))
plt.title('Gaussian blur, σ=5')
Filters

Larger kernels and sigma values will result in a greater smoothing effect and a more blurred image - as you can see, it is possible to over-blur the image.

Removing the background


Eventually we’re going to want to isolate the foreground from the background. In some images, this may be difficult especially if the two are not entirely distinct from each other, or if the background is not a uniform shade. In cases like this, a rolling ball algorithm can be applied. The rolling ball estimates the background intensity of an image by using the pixel values to translate the image into a height map, and then rolling a ball of a given radius across it. The rolling ball algorithm has been published and a figure from the publication describes how it works.

Exercise 7: Rolling ball background intensity

Look at the scikit-image documentation on the rolling ball filter. Load the example image skimage.data.coins and display:

  • the original image
  • image with a rolling ball of radius 100 applied
  • image with a rolling ball of radius 50 applied
  • the image with the radius=50 rolling ball subtracted from it

Compute time can be found using Python’s datetime library:

PYTHON

from skimage.data import coins
from skimage.restoration import rolling_ball

image = coins()
plt.figure(figsize=(10, 8))

plt.subplot(2, 2, 1)
plt.imshow(image, cmap='gray')
plt.title('Original')

plt.subplot(2, 2, 2)
rolling_ball_100 = rolling_ball(image, radius=100)
plt.imshow(rolling_ball_100, cmap='gray')
plt.title('Rolling ball, r=100')

plt.subplot(2, 2, 3)
rolling_ball_50 = rolling_ball(image, radius=50)
plt.imshow(rolling_ball_50, cmap='gray')
plt.title('Rolling ball, r=50')

plt.subplot(2, 2, 4)
plt.imshow(image - rolling_ball_50)
plt.title('Original - rolling ball r50')
Rolling ball

Dogs and logs


Difference of Gaussian (DoG) and Laplacian of Gaussian (LoG) are algorithms that build on the Gaussian filter.

In a Difference of Gaussian, two Gaussian filters are taken of the image, each with a different sigma value. The larger filter is then subtracted from the first to give an image where features are effectively highlighted by an area of high contrast.

In a Laplacian of Gaussian, a Gaussian-filtered image is supplied to a Laplace filter. This eliminates the need to manually select two sigma values as with Difference of Gaussian.

Exercise 8: Dogs and logs

Load a frame from the second channel of skimage.data.cells3d() again:

PYTHON

from skimage.data import cells3d

image = cells3d()[30, 1, :, :]

Display a figure of:

  • The original image
  • Image with a Difference of Gaussians applied with sigma values 2 and 4
  • Image with a Laplacian of Gaussians applied. Note that the Laplace filter linked above does not perform the Gaussian filter for you, so you will need to pass the image through gaussian() first.

PYTHON

from skimage.data import cells3d
from skimage.filters import gaussian, difference_of_gaussians, laplace

image = cells3d()[30, 1, :, :]
plt.figure(figsize=(10, 12))

plt.subplot(2, 2, 1)
plt.imshow(image)
plt.title('Original')

plt.subplot(2, 2, 2)
plt.imshow(difference_of_gaussians(image, 1, 2), cmap='gray')
plt.title('Difference of Gaussians')

plt.subplot(2, 2, 3)
plt.imshow(laplace(gaussian(image, sigma=2)), cmap='gray')
plt.title('Laplacian of Gaussian')
Dogs and logs

Exercise 9: Choosing a filter

Look at the images you produced in exercises 6 and 8, and select one to use in subsequent chapters for thresholding and segmentation!

Key Points

  • There are many ways of smoothing an image
  • Different methods will perform better in different situations

Content from Thresholding and Segmentation


Last updated on 2025-03-24 | Edit this page

Overview

Questions

  • How can I isolate features of interest in my image?
  • How can I uniquely label features once I’ve found them?

Objectives

  • Apply various thresholds an image
  • Quantify features of interest in an image
  • Demonstrate how to deal with edge cases like features stuck together

These exercises will make use of the image you decided to use in exercise 9.

Thresholding


Thresholding consists of converting an image into a binary form where each pixel value is either 0 or 1 depending on whether it exceeds a given threshold.

We can binarise an image with a given threshold by applying the > operator to it. It’s common practice to smooth the image first, e.g. with a Gaussian filter:

PYTHON

import matplotlib.pyplot as plt
from skimage.data import cells3d
from skimage.filters import gaussian

smoothed_image = gaussian(cells3d()[30, 1, :, :])
print(smoothed_image.min(), smoothed_image.max(), smoothed_image.dtype)

plt.subplot(1, 2, 1)
plt.imshow(smoothed_image, cmap='gray')
plt.title('Gaussian filter')

plt.subplot(1, 2, 2)
binary_image = smoothed_image > 0.12
plt.imshow(binary_image)
plt.title('Threshold > 0.12')
Threshold

Applying a set number to an image as a threshold is simple, but it’s not ideal having to guess an appropriate threshold until we get a sensible one. Looking at the histogram helps, but still requires some manual interpretation and human subjectivity:

PYTHON

plt.hist(image.flatten(), bins=256)
Histogram

Fortunately, there are several algorithms for automatically determining an appropriate threshold based on the image’s histogram. Global thresholds work on the entire image, whereas local thresholds work on a window around each pixel - for the moment, we will only discuss global filters.

Exercise 10: Thresholds

Look at the skimage filter docs and find each of the following threshold algorithms:

  • Otsu threshold
  • Triangle threshold

Try running each of these algorithms on your smoothed image from exercise 9, and use the resulting threshold to binarise the image. How do the results differ?

We can display the two threshold algorithms in a figure:

PYTHON

from skimage.filters import threshold_otsu, threshold_triangle

plt.figure(figsize=(10, 8))

plt.subplot(1, 3, 1)
plt.imshow(smoothed_image)
plt.title('Gaussian filter')

# Otsu threshold
plt.subplot(1, 3, 2)
threshold = threshold_otsu(image)
plt.imshow(image > threshold)
plt.title('Otsu threshold')

# Triangle threshold
plt.subplot(1, 3, 3)
threshold = threshold_triangle(image)
plt.imshow(image > threshold)
plt.title('Triangle threshold')
Otsu and Triangle thresholds

Which threshold performs better can depend on the image you have. Otsu thresholding is less sensitive so foreground objects are more likely to be incomplete or have holes in them, whereas triangle thresholding has a greater tendency for close foreground objects to become stuck together.

Cleaning up


Once we have a binary image, we may still need to clean it up further. There may still be extraneous pixels and other bits and pieces left around the regions of interest, or there may be holes in a shape that you need to be solid.

Exercise 11: Erosion, dilation, opening and closing

Look at the skimage morphology docs. Select one of your binary images from exercise 10 and apply each of the following algorithms to the image:

  • Binary erosion
  • Binary dilation
  • Binary opening
  • Binary closing

Note: each of the algorithms above will require a kernel or footprint. For each of these, use a disc-shaped kernel of radius 2:

PYTHON

from skimage.morphology import disk
kernel = disk(2)

PYTHON

from skimage.morphology import binary_erosion, binary_dilation, binary_opening, binary_closing
from skimage.morphology import disk

kernel = disk(2)
plt.figure(figsize=(8, 12))

# your binary image from exercise 10
plt.subplot(3, 2, 1)
plt.imshow(binary_image, cmap='gray')
plt.title('Binary image')

# erosion
plt.subplot(3, 2, 2)
plt.imshow(binary_erosion(binary_image, footprint=kernel), cmap='gray')
plt.title('Erosion')

# dilation
plt.subplot(3, 2, 3)
plt.imshow(binary_dilation(binary_image, footprint=kernel), cmap='gray')
plt.title('Dilation')

# opening
plt.subplot(3, 2, 4)
plt.imshow(binary_opening(binary_image, footprint=kernel), cmap='gray')
plt.title('Opening')

# closing
plt.subplot(3, 2, 5)
plt.imshow(binary_closing(binary_image, footprint=kernel), cmap='gray')
plt.title('Closing')
Image transformations

Eroding has the effect of shrinking all foreground features. Dilating has the opposite effect, potentially resulting in foreground features becoming stuck together.

Opening an image is an erosion followed by a dilation. If your image is grainy with many small artifacts, these will be removed.

Closing is the opposite - dilation followed by erosion. This will have the effect of filling in any holes in your foreground features.

If there are still holes in your binary image, there is also a function in scipy.ndimage that can fix this:

PYTHON

from scipy.ndimage import binary_fill_holes

fill_holes = binary_fill_holes(binary_image)
plt.imshow(fill_holes)
Binary fill holes

Labelling


In some situations, it may be necessary to distinguish between individual features, e.g. if counting cell nuclei. For this, it is necessary to isolate or label individual separated objects in the foreground:

PYTHON

from skimage.morphology import label

labels = label(fill_holes)
plt.imshow(labels, cmap='viridis')
plt.colorbar()

This will result in an image similar to the binary image supplied, but instead of just 1s, each feature will have a unique number assigned to all of its pixels. Displaying this with a colour scale will effectively colour-code each one:

Image labelling

However, we can see that the process is not perfect, as nuclei that are stuck together have been marked as the same feature. In the exercises below, we will explore how to solve this.

Watershed transform


Several transforming or segmentation operations exist for solving situations like the one above with touching objects.

This transformation requires a few preparation steps. First, we need a distance transform of the binary image, where each foreground pixel’s value is its Euclidean distance to the background. skimage doesn’t have a function for this, but scipy does:

PYTHON

from scipy.ndimage import distance_transform_edt

dt = distance_transform_edt(fill_holes)
plt.imshow(dt)
Distance transform

Next, we need the coordinates of all local maxima, i.e. locations that are furthest away from the background. This requires a kernel/footprint, which in this example is a square kernel of 7x7:

PYTHON

from skimage.feature import peak_local_max

coords = peak_local_max(dt, labels=fill_holes)

This gives us a list of coordinates, so we need to convert this back into an labelled image of maxima. This will be mostly blank with an individually labelled dot representing each one:

PYTHON

from numpy import zeros

mask = zeros(dt.shape, dtype=bool)  # create a blank image the same size as the original
mask[tuple(coords.T)] = True  # create a foreground pixel at each coordinate
markers = label(mask)  # uniquely label them
plt.imshow(markers, cmap='viridis')
Local maxima

Note the .T to transpose the maxima coordinates.

At this point depending on your image, you may get clusters of peaks close together rather than one clean, distinct peak per feature. This is especially likely if your features are not circular. Applied to the final watershed step below, this will result in features being split when they shouldn’t be. This can be resolved in several ways:

Smoothing the distance transform first

PYTHON

coords = peak_local_max(gaussian(dt, sigma=4), labels=fill_holes)

Applying a larger footprint to peak_local_max

There are a few different ways to do this - all we need to do is create a kernel of a cerain shape and size, and pass it to peak_local_max:

PYTHON

# numpy
from numpy import ones
coords = peak_local_max(dt, footprint=ones((7,7)), labels=fill_holes)

# skimage < 0.25.0
coords = peak_local_max(dt, footprint=square(7), labels=fill_holes)

# skimage >= 0.25.0
coords = peak_local_max(dt, footprint=footprint_rectangle((7, 7)), labels=fill_holes)

Providing a min_distance argument to peak_local_max

This will prevent the function from identifying peaks too close together:

PYTHON

coords = peak_local_max(dt, min_distance=10, labels=fill_holes)

Masked watershed

Now that we have identified some peaks, we can now put the distance transform together with the local peaks to do the transform:

PYTHON

from skimage.segmentation import watershed

watershed_transform = watershed(-dt, markers)
plt.imshow(watershed_transform, cmap='viridis')
Watershed segmentation

Note the - in front of the distance transform. This effectively converts it from a map of peaks to a map of troughs. We need to do this because of the way watershed transforms work.

The result of the watershed transform above does not indicate the shape of the nuclei. Rather, the algorithm effectively decides which nucleus each pixel of the image most closely corresponds to. We can generate a more meaningful image by multiplying the watershed and the binary image together to create a masked watershed:

PYTHON

masked_watershed = watershed_transform * binary_image
plt.imshow(masked_watershed, cmap='viridis')
Masked watershed

It’s also possible to do the last two steps in one by providing mask=some_binary_image to skimage.segmentation.watershed.

Exercise 12: Image segmentation

  • Load up one of your cleaned up images from exercise 11 and run skimage.morphology.label on the foreground features
  • How many features did skimage find? (hint: try flattening your labelled image and printing the result to the screen)
  • Perform a watershed transform on your binary image as above. Try displaying the result of each stage of the transform.
  • How many features are identified now, after watershed transformation?
  • Does the watershed transformation work perfectly?

Labelling applies a unique number to each feature. This means that if we apply thresholding and labelling to the image, we can flatten the result and cast it to a set to find all the unique label numbers present, and find the highest numbered label to get the number of features:

PYTHON

from skimage.data import cells3d
from skimage.filters import gaussian, threshold_otsu
from skimage.morphology import label

# whatever your chosen cleaned up image is
plt.subplot(1, 2, 1)
plt.imshow(binary_image)
plt.title('Cleaned up image')

# labelled
labels = label(binary_image)
plt.subplot(1, 2, 2)
plt.imshow(labels, cmap='viridis')
plt.title('Labelled')

flat = labels.flatten()
print(set(flat))
print('Features found:', max(flat))

However, we see from the intermediate images that multiple touching nuclei are treated as a single feature.

Here is an example of using distance and watershed transforms to segment touching objects:

PYTHON

from skimage.feature import peak_local_max
from skimage.segmentation import watershed
from scipy.ndimage import distance_transform_edt

dt = distance_transform_edt(binary_image)
plt.subplot(2, 2, 1)
plt.imshow(dt)
plt.title('Distance transform')

coords = peak_local_max(dt, min_distance=10, labels=binary_image)
mask = numpy.zeros(dt.shape, dtype=bool)
mask[tuple(coords.T)] = True
plt.subplot(2, 2, 2)
plt.imshow(mask, cmap='viridis')
plt.title('Markers')

markers = label(mask)
labels = watershed(-dt, markers, mask=binary_image)
plt.subplot(2, 2, 3)
plt.imshow(labels, cmap='viridis')
print('Features found:', max(labels.flatten()))

This will segment more correctly, however the accuracy of the process still depends on the inputs supplied. Irregular objects with holes in them will be harder to segment than even, solid, rounded shapes.

Key Points

  • Thresholding, labelling and feature isolation are all forms of segmentation
  • Different thresholding algorithms can perform differently in different situations
  • Quantitative image analysis often requires us to individually label features
  • Processing such as watershed transforms can solve cases where features of interest are very close or stuck together

Content from Measurements


Last updated on 2025-03-21 | Edit this page

Overview

Questions

  • How can I make useful measurements in my image?

Objectives

  • Find the mean pixel intensity in each nucleus
  • Find the area of each nucleus

Having isolated individual nuclei in the example image in the previous chapter, we can make some observations about the image features. The exercises here will require:

  • Your masked watershed transformation from the previous chapter
  • Your original smoothed image from exercise 9

Processing one nucleus at a time

Remember that each nucleus is labelled with a unique number, so we use masking to select a single nucleus. We can construct a binary mask from a labelled image using numpy’s where function:

PYTHON

from numpy import where

nucleus_1 = where(masked_watershed == 1, 1, 0)
nucleus_2 = where(masked_watershed == 2, 1, 0)

Here, each pixel will be set to either 1 or 0 depending on whether its value matches the label we’re looking for, effectively creating a binary mask for each nucleus.

Exercise 13: Quantifying intensities

For each labelled nucleus:

  • Construct a binary mask matching the label number
  • Use this to isolate the individual nucleus out of your original smoothed image
    • Hint: try looking at the numpy.where docs to see the different ways it can be used.
  • Find the mean pixel value for the nucleus

First, we need to know how many features we’re looping over. We can get that from the numbers that the labelling process provides - if your labelled image contains the labels 1, 2 and 3, then you have 3 features.

Next, we need to generate a mask for that feature using numpy.where, that will allow us to mask out the original image - in this case, either a corresponding pixel value from the original image or None. We can then take the result of that, select all pixels that are not None and use numpy.mean.

PYTHON

plt.figure(figsize=(12, 8))
labels = sorted(set(masked_watershed.flatten()))

for i in labels[:6]:  # just showing the first 6
    mask = numpy.where(masked_watershed == i, 1, 0)
    nucleus = numpy.where(mask == 1, image, 0)
    plt.subplot(2, 3, i + 1)  # subplot counts from 1 rather than 0
    plt.imshow(nucleus)
    pixels = [p for p in nucleus.flatten() if p]  # remove the background pixels
    print('Feature ' + str(i) + ':', numpy.mean(pixels))
Mean intensities

A few notes:

  • Note how the first feature labelled as 0 corresponds to the background
  • Remember that in subplot, we need to offset the subplot number by +1, since it counts from 1
  • In the first numpy.where above, we supply two single numbers, 1 and 0, giving us a binary mask. In the next one, we provide an n-dimensional array instead, causing it to instead select the value of the corresponding pixel.

Key Points

  • Binary multiplication and numpy.where are powerful ways of combining images and masks
  • The numbers assigned during labelling can be used to select and process one feature at a time

Content from Introduction to Napari


Last updated on 2025-03-18 | Edit this page

Overview

Questions

  • How can I view images and metadata in Napari?

Objectives

  • Open and explore some images and compare the process of doing the same process in Python

Starting Napari


With Napari installed and its Conda environment active, run:

SH

napari

This will start up Napari in a new window. This may take about a minute, especially the first time.

Looking around


In the ‘File’ menu, go to ‘Open Sample -> napari builtins -> Cells (3D+2Ch)’. Looking at the sidebar, Napari has opened two layers - ‘membrane’ and ‘nuclei’. At the top of the sidebar there are options for displaying layers in different colours and opacity, and the eye symbol on the layer can be checked or unchecked to show/hide it:

Napari layers

Napari uses layers to represent each stage of processing an image - we run tools to create new processed layers based off of a previous one. In this example, the image consists of two channels, each of which has been loaded as a layer.

In the case of z-stack or time series images such as this one, there is also a slider at the bottom for scrolling through the layers.

Exercise 14: Napari tools

With ‘Cells (3D + 2Ch)’ open, go to ‘Tools -> Filtering / Noise Removal’ and look at the options available. What are these options? What happens if you select one and try running it on a layer?

The options here consist of smoothing filters, several of which will be familiar from previous chapters. In some cases there are multiple versions of a filter (e.g. there are three Gaussian filters), each of which uses a different Python library.

Selecting one of these gives a sidebar letting us choose how to run it, including which layer to use, plus any extra parameters it may take, such as sigma values for Gaussian filters:

Gaussian filter controls
Gaussian filter

Loading real-world images


The example images work nicely with Napari, but sometimes it may be more difficult to coax Napari into loading an image correctly. In ‘File’, select ‘Open File(s)…’ and load up FluorescentCells_3channel.tif. Doing this, we find the results aren’t quite what we expect - the three channels have been loaded as if they’re a z-stack.

Fortunately, Napari includes a Python console that gives us very fine-grained control of the user interface, and will let us load this image correctly. In the bottom left, use the ’>_’ button to open a terminal. This activates a Python session where Napari is represented as an object called viewer:

Napari Python console
  • All the loaded layers are in a list called viewer.layers
  • Within a layer, the image data is an attribute called data

In this console, we can create new layers with viewer.add_image. First, let’s look at the image’s shape:

PYTHON

from skimage.io import imread
image = imread('path/to/FluorescentCells_3channel.tif')
image.shape
# returns: (512, 512, 3)

From this, we can see that the colour channels are the third axis. Using this, We can now load each channel as a new layer:

PYTHON

viewer.add_image(image[:, :, 0], name='FluorescentCells_ch0', colormap='red')
viewer.add_image(image[:, :, 1], name='FluorescentCells_ch1', colormap='green', blending='additive')
viewer.add_image(image[:, :, 2], name='FluorescentCells_ch2', colormap='blue', blending='additive')

The blending='additive' option prevents layers on top from obscuring layers below it so that we can view multiple layers at once. We also use colormap to colour the layers so we can distinguish them from each other.

viewer.add_image is also able to load all these layers in one if we tell it which axis represents the colour channels:

PYTHON

viewer.add_image(image, name='FluorescentCells', channel_axis=2)

Exercise 15: Real world images

  • Go to ‘File’ -> ‘Open File(s)…’ and load the image ‘confocal-series_zstack.tif’. How does it look? Can you view each channel simultaneously?
  • Now try loading the same image in the console as above, looking at the image’s shape and passing the channel axis to viewer_add_image. axis looks like the channels. How has the fourth axis been represented?

Loading the image from the menu, we can see some results but it’s all in one layer and the channel axis is represented as a slider, meaning we can’t view the two at the same time.

As an alternative, we can load the image in the Python console:

PYTHON

image = imread('path/to/confocal-series_zstack.tif')
image.shape
# returns: (25, 2, 400, 400)

The second number, i.e. image.shape[1] looks like the channel axis. We can then provide this when loading the image:

PYTHON

viewer.add_image(image, name='FluorescentCells', channel_axis=1)

Napari will load the image as two coloured layers, with the Z axis represented via the slider at the bottom.

Exercise 16: Image segmentation in the Napari console

This exercise will put together everything that has been covered in previous chapters, and apply the same steps in the Napari console.

Look at the ‘nuclei’ layer of ‘Cells (3D + 2Ch)’. In the Python console, try doing the same smoothing, binarisation, segmentation and labelling that you performed throughout exercises 6-12, except displaying each step as a new layer in the Napari viewer.

Remember that instead of matplotlib.pyplot.imshow(), you’ll need to use viewer.add_image()

PYTHON

import skimage

image = skimage.data.cells3d()
image.shape
# returns: (60, 2, 256, 256)

frame = image[30, 1, :, :]
smoothed_image = skimage.filters.gaussian(frame, sigma=1)
viewer.add_image(smoothed_image, name='smoothed')

threshold = skimage.filters.threshold_otsu(smoothed_image)
binary_image = smoothed_image > threshold
viewer.add_image(binary_image, name='Otsu threshold')

import scipy.ndimage
fill_holes = scipy.ndimage.binary_fill_holes(binary_image)
viewer.add_image(fill_holes, name='fill_holes')

distance_transform = scipy.ndimage.distance_transform_edt(fill_holes)
viewer.add_image(distance_transform, name='distance_transform')

coords = skimage.feature.peak_local_max(skimage.filters.gaussian(distance_transform, sigma=4), labels=fill_holes)

import numpy
mask = numpy.zeros(distance_transform.shape, dtype=bool)
mask[tuple(coords.T)] = True
markers = skimage.morphology.label(mask)
viewer.add_image(markers, name='peaks')

watershed_transform = skimage.segmentation.watershed(-distance_transform, markers)
viewer.add_image(watershed_transform, name='watershed')

masked_watershed = watershed_transform * fill_holes
viewer.add_image(masked_watershed, name='masked_watershed')

Image segmentation with Napari tools


Aside from running analyses in the Python console, Napari has a variety of processing steps that can be browsed under ‘Tools’.

For example, we can go to ‘Open Sample’ -> ‘napari builtins’ and -> ‘Binary Blobs’. Browse the tools available and run:

  • Otsu threshold
  • Split touching objects (nsbatwm). Try experimenting with different sigma values until you get some good cell separations.
  • Connected component labelling (scikit-image, nsbatwm)

Note: The Otsu threshold shouldn’t be required since the test image is already a binary, but the ‘Split touching objects’ Napari tool isn’t able to recognise the original as a binary image. This is because Napari expects the values to be 0/1, whereas the test image uses the values True/False. Thresholding the image again converts it to the right type.

Exercise 17: Exporting code

This exercise will see how the processing that we have built up with Napari tools can be exported back to Python code that we can then check into version control and integrate into a workflow to help automate the processing of many images.

Under ‘Plugins’, open the Napari Assistant and find the ‘Export Python code’ tool. Use this to generate a Python script that will perform the processing that you just performed above on the Binary Blobs image. Does the script look usable immediately without modification?

There are several options to export in different ways, including to a Jupyter notebook. The simplest way is probably to copy to clipboard - this can be pasted into a text editor or IDE:

PYTHON

from skimage.io import imread
import napari_segment_blobs_and_things_with_membranes as nsbatwm  # version 0.3.6
import napari
if 'viewer' not in globals():
    viewer = napari.Viewer()

image0_bb = viewer.layers['binary_blobs'].data

# threshold otsu

image1_to = nsbatwm.threshold_otsu(image0_bb)
viewer.add_labels(image1_to, name='Result of threshold_otsu')

# split touching objects

image2_sto = nsbatwm.split_touching_objects(image1_to, 5.6)
viewer.add_labels(image2_sto, name='Result of split_touching_objects')

# connected component labeling

image3_ccl = nsbatwm.connected_component_labeling(image2_sto, False)
viewer.add_labels(image3_ccl, name='Result of connected_component_labeling')

Looking at it, we can see that it’s not immediately usable but it’s pretty close. Exported scripts do not make any reference to what image file was loaded, and there is no code at the end for writing the resuults to a file, so this is what you would need to add to get it to produce tangible output.

Key Points

  • Napari works with layers, each of which represents an image channel
  • Napari doesn’t always know how to load a multi-channel image from the GUI
  • We can use the Python console to perform custom operations that can’t be done in the GUI
  • Most operations we performed earlier in Python can also be done in Napari, either graphically or in the terminal