1 / 19

Today’s Topic

Today’s Topic. (Slight ) correction to last week’s lecture Getting z-resolution—confocal detection and deconvolution algorithms. (Lab 1) Polarization assays—important for k 2 and for binding assays (Lab 2). Time (nsec). What is fluorescence? (correction).

aiden
Télécharger la présentation

Today’s Topic

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Today’s Topic (Slight ) correction to last week’s lecture Getting z-resolution—confocal detection and deconvolution algorithms. (Lab 1) Polarization assays—important for k2 and for binding assays (Lab 2)

  2. Time (nsec) What is fluorescence? (correction) Shine light in, gets absorbed, reemits at longer wavelength Thermal Relaxation [psec] Stokes Shift (10-100 nm) Light In Excitation Spectra Emission Spectra Light Out Absorption [ftsec] Old diagram was wrong Thermal relaxation [psec] Fluorescence & + non-radiative [nsec] k k= kf +kn.r. Fluorescence -t/to y= e to = 1/k= 1/(kf +kn.r.) Photobleaching Important: Dye emits 105 107 photons, then dies!

  3. (De-)Convolution

  4. (De-)Convolution

  5. (De-)Convolution

  6. Getting 3DConvolution (pinhole) vs. Widefield Deconvolution

  7. (Last time) Convolution (pinhole) microscopy Way to improve the z-resolution by collecting less out-of-focus fluorescence, via a pinhole and scanning excitation with focused light. Great especially for thick samples, but takes time, complicated optics and requires fluorophores that are very stable. (why?) Focused Light creates fluorescence which gets to detector Fluorescence (dark-field) Confocal microscopy prevents out-of-focus blur from being detected by placing a pinhole aperture between the objective and the detector, through which only in-focus light rays can pass. Light mostly gets rejected http://micro.magnet.fsu.edu/primer/digitalimaging/deconvolution/deconintro.html

  8. What if you let light go in (no pinhole) and don’t scan—i.e. use wide-field excitation. Can you mathematically discriminate against out-of-focus light? Deconvolution For each focal plane in the specimen, a corresponding image plane is recorded by the detector and subsequently stored in a data analysis computer. During deconvolution analysis, the entire series of optical sections is analyzed to create a three-dimensional montage. http://micro.magnet.fsu.edu/primer/digitalimaging/deconvolution/deconintro.html

  9. Deconvolution: Deconvolution is a computationally intensive image processing technique that is being increasingly utilized for improving the contrast and resolution of digital images captured in the microscope. The foundations are based upon a suite of methods that are designed to remove or reverse the blurring present in microscope images…. In contrast, widefield microscopy allows blurred light to reach the detector (or image sensor), but deconvolution techniques are then applied to the resulting images either to subtract blurred light or to reassign it back to a source. Even if you have a point emitter (at z=0), you will see a PSF like cone. By knowing the (mathematical) transfer function, can you do better? Called deconvolution techniques. Common technique: take a series of z-axis and then unscramble them. http://micro.magnet.fsu.edu/primer/digitalimaging/deconvolution/deconintro.html

  10. Deconvolution: 2 TechniquesDeblurring and image restoration DeblurringAlgorithmsare fundamentally two-dimensiona: they subtract away the average of the nearest neighbors in a 3D stack. For example, the nearest-neighbor algorithm operates on the plane z by blurring the neighboring planes (z + 1 and z - 1, using a digital blurring filter), then subtracting the blurred planes from the z plane. In contrast, image restoration algorithms are properly termed "three-dimensional" because they operate simultaneously on every pixel in a three-dimensional image stack. Instead of subtracting blur, they attempt to reassign blurred light to the proper in-focus location. • +Computationally simple. • Add noise, reduce signal • Sometimes distort image http://micro.magnet.fsu.edu/primer/digitalimaging/deconvolution/deconalgorithms.html

  11. Deconvolution Image Restoration Instead of subtracting blur (as deblurring methods do), Image Restoration Algorithmsattempt to reassign blurred light to the proper in-focus location. This is performed by reversing the convolution operation inherent in the imaging system. If the imaging system is modeled as a convolution of the object with the point spread function, then a deconvolution of the raw image should restore the object. However, never know PSF perfectly. Varies from point-to-point, especially as a function of z and by color. You guess or take some average.

  12. Deconvolution algorithms and Fourier Transforms An advantage of this formulation is that convolution operations on large matrices (such as a three-dimensional image stack) can be computed very simply using the mathematical technique of Fourier transformation. If the image and point spread function are transformed into Fourier space, the convolution of the image by the point spread function can be computed simply by multiplying their Fourier transforms. The resulting Fourier image can then be back-transformed into real three-dimensional coordinates.

  13. Deconvolution Algorithms A three-dimensional image stack containing 70 optical sections, and recorded at intervals of 0.2 micrometers through a single XLK2 cell. A widefield imaging system, equipped with a high numerical aperture (1.40) oil immersion objective, was employed to acquire the images. The Original Datais a single focal plane taken from the three-dimensional stack, before the application of any data processing. Deblurring by a nearest-neighbor algorithm produced the result shown in the image labeled Nearest Neighbor (Figure 2(b)). The third image (Figure 2(c), Restored) illustrates the result of restoration by a commercially marketed constrained iterative deconvolution software product. Both deblurring and restoration improve contrast, but the signal-to-noise ratio is significantly lower in the deblurred image than in the restored image. The scale bar in Figure 2(c) represents a length of 2 micrometers, and the arrow (Figure 2(a))designates the position of the line plot presented in Figure 4.

  14. Contrast: deblur & image restoration The graphical plot presented in Figure 4 represents the pixel brightness values along a horizontal line traversing the cell illustrated in Figure 2 (the line position is shown by the arrow in Figure 2(a)). Original data = green line, the deblurred image data = blue line, and the restored image data = the red line. Deblurringcauses a significant loss of pixel intensity over the entire image, whereas restoration results in a gain of intensity in areas of specimen detail. A similar loss of image intensity as that seen with the deblurring method occurs with the application of any two-dimensional filter.

  15. Estimating the object First, an initial estimate of the object is performed, which is usually the raw image itself. The estimate is then convolved with the point spread function, and the resulting "blurred estimate" is compared with the original raw image. This comparison is employed to compute an error criterion that represents how similar the blurred estimate is to the raw image. Often referred to as a figure of merit, the error criterion is then utilized to alter the estimate in such a way that the error is reduced. A new iteration then takes place: The entire process is repeated until the error criterion is minimized or reaches a defined threshold. The final restored image is the object estimate at the last iteration.

  16. Blind Deconvolutionan image reconstruction technique:object and PSF are estimated The algorithm was developed by altering the maximum likelihood estimation procedure so that not only the object, but also the point spread function is estimated. Another family of iterative algorithms uses probabilistic error criteria borrowed from statistical theory. Likelihood, a reverse variation of probability, is employed in the maximum likelihood estimation (MLE) Using this approach, an initial estimate of the object is made and the estimate is then convolved with a theoretical point spread function calculated from optical parameters of the imaging system. The resulting blurred estimate is compared with the raw image, a correction is computed, and this correction is employed to generate a new estimate, as described above. This same correction is also applied to the point spread function, generating a new point spread function estimate. In further iterations, the point spread function estimate and the object estimate are updated together.

  17. Different Deconvolution Algorithms The original (raw) image is illustrated in Figure 3(a). The results of deblurring by a nearest neighbor algorithm appear in Figure 3(b), with processing parameters set for 95 percent haze removal. The same image slice is illustrated after deconvolution by an inverse (Wiener) filter (Figure 3(c)), and by iterative blind deconvolution (Figure 3(d)), incorporating an adaptive point spread function method. http://micro.magnet.fsu.edu/primer/digitalimaging/deconvolution/deconalgorithms.html

  18. Slice #36 from wide-field fluorescence Z-stack Deconvolution Deconvoluted (using theoretical PSF) Wide-Field Fluorescence slice 36 DECONVOLUTED MIN PROJECTION DECONVOLUTED MAX PROJECTION

  19. The End

More Related