1 / 46

Computer Graphics & Image Processing Lecture 2 Digital Image Fundamentals

Computer Graphics & Image Processing Lecture 2 Digital Image Fundamentals. ALI JAVED Lecturer SOFTWARE ENGINEERING DEPARTMENT U.E.T TAXILA Email:: alijaved@uettaxila.edu.pk Office Room #:: 7. Visual Perception. •How images are formed in the eye ?

tanika
Télécharger la présentation

Computer Graphics & Image Processing Lecture 2 Digital Image Fundamentals

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Computer Graphics & Image Processing Lecture 2Digital Image Fundamentals

  2. ALI JAVED Lecturer SOFTWARE ENGINEERING DEPARTMENT U.E.T TAXILA Email:: alijaved@uettaxila.edu.pk Office Room #:: 7

  3. Visual Perception •How images are formed in the eye ? •Eye’s physical limitation ? •Human visual interpretation of images ?

  4. Structure of human eyes

  5. Structure of human eyes • Three membranes enclose the eye: • – Cornea and sclera • • Cornea is a tough, transparent tissue cover the anterior surface of the eye. • •Sclera is a opaque membrane enclose the remainder of the optic globe. • –Choroid • • A network of blood vessels for eye nutrition • • At its anterior extreme, it is divided into the ciliary body and iris diaphragm. • • The central opening (the pupil) varies in diameter from 2 to 8 • mm. • –Retina • Retina lines the insides of the wall’s interior portion with two classes of receptors:

  6. Structure of human eyes – Cones: (Red 65%, Green 33%,Blue 2%) • 6 – 7 millions located primarily in the central portion of the retina • Highly sensitive to color • Photopic or bright-light vision – Rods • 75- 150 millions distributed over the retinal surface. • Not involved in color vision and sensitive to low illumination • Scoptopic or dim vision

  7. Structure of human eyes The cones are most dense in the center of retina. Density of cones in the area of fovea is 150,000 element/mm2 The number of cones in fovea is 337,000 elements.

  8. Image Formation in the Eyes The distance between the center of the lens and the retina (focal length) varies from 17mm to 14mm. The shape of lens is controlled by the tension of fibers of the ciliary body. The retinal image is reflected primarily in the area of fovea. Perception = excitation of light receptors, which transform radiant energy into impulses that are ultimately decoded by the brain.

  9. Image Formation in the Eyes

  10. Brightness Adaptation & Discrimination The range of light intensity levels to which the human visual system can adapt is enormous - on the order of 10^10. • The subjective brightness is a logarithmic function of light intensity incident on the eye. • Brightness adaptation. – The current sensitivity level it can discriminate simultaneously is rather small compared with the total adaptation range – Brightness adaptation level: the current sensitive level of the visual system.

  11. Brightness Adaptation & Discrimination

  12. Brightness Adaptation & Discrimination

  13. Brightness Adaptation & Discrimination •The dIc is the increment of illumination discriminable 50% of the time with the background illumination I. • The quantity dIc/I is called the Weber ratio. • The smaller dIc/I means that a small percentage change in intensity is discriminable – good brightness discrimination • If the background illumination is constant, the intensity of object is allowed to vary incrementally from never perceived to always being perceived. • Typically the observer can discern a totally from one to two dozens different intensity changes. – The number of gray level for digital image – Contouring Effect: Not sufficient no. of gray levels

  14. Brightness Adaptation & Discrimination Perceived brightness is not a simple function of intensity, rather it is log of intensity

  15. Brightness Adaptation & Discrimination – A region’s perceived brightness does not simply depend on its intensity (fig. 2.8) – Simultaneous contrast.

  16. Light and EM Spectrum

  17. Light and EM Spectrum • Light is a particular type of EM radiation that can be seen by human eye. • EM waves are massless particles each traveling in a wavelike pattern and moving at a speed of light. • We can specify waves through frequency and wavelength. • The colors that human perceive in an object are determined by the nature of the light reflected from the object. • For example green objects reflect light with wavelengths primarily in the 500 to 570nm range while absorbing most of the energy at other wavelengths.

  18. Light and EM Spectrum • Achromatic Light Light that is void of color is called achromatic or monochromatic light. The only attribute of such light is its intensity. The term gray level generally is used to describe monochromatic intensity because it ranges from black to grays and finally to white • Chromatic light spans EM spectrum from 0.43 um (violet) to 0.79 um ( red). Three basic quantities are used to describe the quality of a chromatic light source – Radiance – Luminance – Brightness

  19. Light and EM Spectrum • Radiance The total amount of energy that flows from the light source Measured in Watts(W) • Luminance Gives a measure of the amount of energy an observer perceives from the a light source. Measured in Lumens(lm) For example light emitted from a source operating in a far infrared region of the spectrum could have significant energy (radiance) but an observer would hardly perceive it; its luminance would be hardly zero • Brightness Subjective descriptor of light perception that is practically impossible to measure

  20. Image Acquisition

  21. Image Acquisition

  22. Image Acquisition using point sensor • Specify the location of vertical and horizontal motors • Sense the light reflection • Voltage waveform will be received (Analog signal) • Convert this analog signal into digital signal through sampling and quantization • - Apply Sampling to digitize coordinate values • - Apply Quantization to digitize amplitude values • Store the digitized value in memory

  23. Image Acquisition using Line sensor • Specify the location of vertical motor • Sense the light reflection • Voltage waveform will be received (Analog signal) • Convert this analog signal into digital signal through sampling and quantization • - Apply Sampling to digitize coordinate values • - Apply Quantization to digitize amplitude values • Store the digitized value in memory

  24. Image Acquisition using Array sensor • Specify the location of vertical and horizontal motors • Sense the light reflection • Voltage waveform will be received (Analog signal) • Convert this analog signal into digital signal through sampling and quantization • - Apply Sampling to digitize coordinate values • - Apply Quantization to digitize amplitude values • Store the digitized value in memory

  25. Image Formation Model

  26. Sampling and Quantization Sampling: Digitization of the spatial coordinates (x,y) Quantization: Digitization in amplitude (also called gray-level quantization) 8 bit quantization: 28 =256 gray levels (0: black, 255: white) Binary (1 bit quantization):2 gray levels (0: black, 1: white) Commonly used number of samples (resolution) Digital still cameras: 640x480, 1024x1024, up to 4064 x 2704 Digital video cameras: 640x480 at 30 frames/second

  27. Sampling and Quantization

  28. Representing Digital Images Digital image is expressed as

  29. Spatial and Gray level Resolution

  30. Spatial and Gray level Resolution

  31. Spatial and Gray level Resolution

  32. Spatial and Gray level Resolution

  33. Spatial and Gray level Resolution

  34. Spatial and Gray level Resolution

  35. Spatial and Gray level Resolution

  36. Aliasing and Moire Patterns • Aliasing: • – If a function is under-sampled then a phenomena called aliasing corrupts the sampled image. • – A signal sampled too slowly is misrepresented by the samples, “ high spatial frequency components of the original signal appear as low spatial frequency components in the sampled signal (an effect known as aliasing. • –The corruption is in the form of additional frequency components being introduced into the sampled function. These are called Aliased frequencies. • –The principal approach for reducing the aliasing effects on an image is to reduce its high frequency components by blurring the image • •Moire Patterns • –Moire Pattern is caused by a break-up of the periodicity, i.e., images are scanned from a printed page, which consists of periodic ink dots.

  37. Zooming and Shrinking Zooming: –Create a new pixel locations –Assign a gray-levels to those new locations •Nearest neighbor interpolation –Pixel replication –Bilinear interpolation using four nearest neighbors –Higher-order non-linear interpolation: using more neighbors for interpolation •Shrinking: –Direct shrinking causes aliasing –Expansion then Shrinking: blurring the image before shrinking it and reduce aliasing.

  38. Zooming and Shrinking

  39. Basic Relationships between Pixels

  40. Basic Relationships between Pixels

  41. Basic Relationships between Pixels

  42. Basic Relationships between Pixels

  43. Basic Relationships between Pixels

  44. Distance Measures Chessboard distance between p and q:

  45. Image operation on a pixel basis

  46. Any question

More Related