Saturday, November 26, 2011

Practice questions for second internal exam

Digital Image Processing
Practice questions for second internal exam

Unit-5 : Colour Image Processing
  1. List the various color models that are used today. Justify their need.
  2. Explain the RGB, CMY, CMYK, HIS and YCrCb color models in detail.
  3. What is the need for HIS model ?  Explain with neat sketches how this model is related to the RGB model. Give suitable conversion formulae.
  4. Provide the conversion formulae for representing a color in one model as a color in a different model.
  5. Distinguish between pseudo color and full color.
  6. Discuss the different methods used in pseudo color processing.
  7. Discuss the different methods used in full color processing.
Unit -6 : Image restoration and reconstruction
  1. How is image restoration different from image enhancement ?
  2. With the help of an image degradation model explain the important causes that lead to image degradation. How can they be removed ?
  3. Using detailed mathematical formulae and graphical representation explain in detail the various noise models that are used during image restoration.
  4. How do you determine which noise model is applicable for a given situation ?
  5. Explain the different types of spatial filters used to remove noise from a degraded image. Comment on their suitability.
  6.  How are degradation functions estimated  ?  Discuss in detail.
  7. Discuss the various inverse filtering schemes and bring out their relative merits and demerits.
Unit-7 : Image segmentation
  1. What is segmentation ? Briefly explain the basis  on which it is done.
  2. When is segmentation successful ? Explain with examples.
  3. What is the role played by the first and second derivatives in detecting an edge ? Explain the relative merits and demerits of both.
  4. Explain in detail how point, line and edge detection is done.
  5. Classify edges using simple sketches. List the steps performed in edge detection ?
  6. Discuss the role of masks in edge detection.
  7. How does thresholding help in edge detection ? What is hysteresis thresholding ?
  8. Explain the Marr-Hildreth edge detection process in detail.
  9. Explain the Canny edge detection process in detail.
Unit – 8 : Image Compression
  1. What are the objectives of image compression ?
  2. What are the different means by which image compression can be achieved ?
  3. Give examples of different types of redundancies that may exist in images.
  4. Explain the basic concepts such as information, data, compression, mapping, channel coding, entropy, quantization , source coding, average length of a code, efficiency based on the foundation of information theory. Use formulae and examples wherever necessary.
  5. What schemes will you propose to remove coding redundancy /spatial redundancy/ttemporal redundancy/irrelevant information ? Give examples.
  6. Discuss the fidelity criteria to quantify loss of information. What care will you exercise while analyzing the results ?
  7. With a neat sketch explain the image compression and decompression process. Explain the function (with examples) of each block in the system.
  8. Problems based on the following compression algorithms : Huffman coding, Arithmetic coding, Lempel-Ziv-Welch coding.
  9. Explain in detail the theory behind : Golomb coding, Block Transform Coding, JPEG coding , Predictive coding.
-----------------------------------------------------------------------------------------------------------------------------------------------------------------

Tuesday, October 11, 2011

Image Compression


Image Compression

1.       Objective :  Reduce the amount of data required to represent an image.
2.       It is both an art and a science.
3.       It is useful and commercially successful technology in DIP.
4.       Need: Reduce the memory and help in increased data transfer per second.
5.       Benefits of compression :
a.       A 2 hour movie stored without compression  in a DVD requires 27 dual layer DVDs of 8.5GB capacity.
b.       The time required to transmit a small 128x128x24bit full-colour image over a 56kbps or 12Mbps(broadband) is from 7.0 to 0.03 secs.
6.       Compression can reduce transmission time by a factor of 2 to 100 or more.
7.       Other areas : televideo conferencing, remote sensing, document and medical imaging, FAX.
8.       We study only the most frequently used compression techniques and some industry standards that make them useful.
9.       Data compression is the process of reducing the amount of data required to represent  a given quantity of information.
10.   Data and information are not the same thing.
11.   Data are the means by which the information is conveyed.
12.   Various amounts of data can be used to represent the same amount of information.
13.   Redundant data : representations that contain irrelevant and/or repeated information.
14.   Relative data redundancy is obtained from compression ration C ( = b/b’ ) as R = 1 – 1/C where b and b’ represent the number of bits of the same picture before and after compression.
15.   Data redundancy in 2D intensity arrays is :
a.       Coding redundancy : The 8-bit codes that are used to represent the intensities in most 2-D intensity arrays contain more bits than are needed to represent the intensities.
b.       Spatial and temporal redundancy : Pixels of most 2D arrays are spatially correlated. Also in a video, the pixels are temporally correlated.
c.       Irrelevant information :  Most 2D arrays contain information that is ignored by the human visual system.
16.   Coding redundancy is present when the codes assigned to a set of events (such as intensity values) do not take full advantage of the probability of the events.
17.   Coding redundancy is almost always present when the intensities of an image are represented using a natural binary code. The reason is that most images are composed of objects that have a regular and somewhat predictable morphology (shape) and reflectance, and are sampled so that the objects being depicted are much larger than the picture elements.
18.   The natural consequence is that for most image, certain intensities are more probable than others.
19.   A natural binary encoding assigns the same number of bits to both the mot and the least probable values, failing to minimize and resulting in coding redundancy.
20.   The compression results from assigning fewer bits to the more probable intensity values than to the less probable one. In the resulting ‘variable length code’ , the image’s most probable intensity is assigned the 1-bit code word while the least probable occurring intensity is assigned the 3-bit code word. Note that the best fixed-length code that can be assigned to the intensities of the image in Eg 8-1 is the natural 2-bit counting sequence but the resulting compression is 4:1 not 4.42:1 which is about 10% less than the 4.42:1 compression of the variable length code.
21.   Spatial and temporal redundancy :  When an image cannot be compressed by variable length coding alone and when all intensity levels have equal probability but when observations reveal spatial redundancy that can be eliminated by representing the image as a sequency of run-length pairs where each run-length pair specifies the start of a new intensity and the number of  consecutive pixels that have that intensity.
22.   In most images , pixels are correlated spatially (in both x and y) and in time(t) when the image is part of a video sequence.
23.   To reduce the redundancy associated with spatially and temporally correlated pixels , a 2-D intensity array must be transformed into a more efficient but usually ‘non-visual’ representation.
24.   When 2D intensity array is converted to run-lengths or the differences between adjacent pixels is used, the transformation is called mapping.
25.   A mapping is reversible if the pixels of the original 2D intensity array can be constructed without error from the transformed data set; otherwise the mapping is said to be irreversible.
26.   Irrelevant information : Compression by removing ‘superfluous’ data from the set. Eg. A homogeneous gray image can be represented by its average intensity alone – a single 8-bit value.
27.   Whether or not this information should be preserved is application dependent. If the information is important (like digital X-ray archive), it should not be omitted; otherwise , the information is redundant and can be excluded for the sake of compression performance.
28.   How to decide the bits that are actually needed to represent the information in an image ? Information theory helps.
29.   Information theory : Generation of information can be modeled as a probabilistic process that can be measured in a manner that agrees with intuition.


30.   A random event E with probability P(E) contains I(E) units of information where I(E) = -log P(E).
31.   If an event occurs always P(E) = 1 and hence no information is attached to it.
32.   The base of the logarithm decides the units used to measure the information. If base = 2 and P(E) = 0.5, I(E) = 1 bit. Meaning : 1 bit is the amount of information conveyed when one of two possible equally likely events occurs.
33.   The entropy of the intensity source  H~ =  negative summation of Pr(rk)log2Pr(rk) from k=0 to L-1.
34.   The amount of entropy and thus information in an image is far from intuitive.
35.   Shannon’s first theorem or noiseless coding theorem :
Lim as n tends to infinity  of [Lavg,n / n] = H where Lavg,n is the average number of code symbols required to represent all n-symbol groups.
36.   Fidelity criteria: Removal of “irrelevant visual”  information involves a loss of real or quantitative image information. How to quantify this loss ? There are two ways
a.       Objective fidelity criteria
b.       Subjective fidelity criteria
37.   Objective fidelity criteria : The information loss can be expressed as a mathematical function of the input and output of a compression process. Eg. RMS error between two images. This is a simple and convenient way to evaluate information loss but not meaningful for human.
38.   Subjective fidelity criteria :  Measuring image quality by subjective evaluations of people by presenting a decompressed image to a cross section of viewers and averaging their evaluation.
39.   Subjective evaluations can be as follows : { -3,-2,-1,0,1,2,3} for {much worse, worse, slightly worse, same, slightly better, better, much better} respectively.
40.   Care must be taken while choosing the results of the above two criteria. Because, a low rms error may also be due to an artificially generated image.
41.   Image compression models : The model of a image compression and decompression consists of two distinct functional components : an encoder and a decoder. This can be done using hardware and/or software.
42.   Codec :  A device/program capable of both encoding and decoding.
43.   Compression process  has three independent operations : mapping , quantizing and symbol coding.
a.       Mapping is a reversible process and transforms f(x,y) to a non-visual format designed to reduce spatial and temporal redundancy.
b.       Quantising is an irreversible process to reduce accuracy of the mapper output in accordance with a preestablished fidelity criterion. The goal is to keep irrelevant information out of the compressed representation.
c.       Symbol coder  is  reversible and generates a fixed-length or variable-length code to represent the quantiser output and maps the output according to the code.
44.   Shortest code words are assigned the most frequently occurring quantizer output values – thus minimizing coding redundancy.
45.   The decoder contains symbol decoder and inverse mapper. Obviously there is no de-quantiser as it is an irreversible process.
46.    In video applications, decoded output frames are maintained in an internal frame store and used to reinsert the temporal redundancy that was removed at the encoder.
47.   Image formats, containers and compression standards :
a.       Image file format is a standard way to organize and store image data. It defines how the data is arranged and the type of compression – if any – that is used.
b.       Image container  is similar to a file format but handles multiple types of image data.
c.       Image compression standards define procedures for compressing and decompressing images.
48.   Standards  for continuous tone still image : JPEG, JPEG-LS, JPEG 2000, BMP, GIF, PDF, PNG, TIFF.
49.   VIDEO Standards : DV, H.261, H.262, H.263, H.264, MPEG-1, MPEG-2, MPEG-4, MPEG-4 AVC, AVS, HDV, M-JPEG, QUICK-TIME, VC-1, WMV9.
50.   BASIC COMPRESSION STANDARDS :  
a.       Huffman coding
b.       Golomb coding
c.       Arithmetic coding
d.       Lempel-Ziv-Welch (LZW) coding
e.       Run-Length coding
f.        Bit-plane coding
g.       Block Transform coding
h.       Predictive coding  :- (i) lossless  (ii) lossy
i.        Wavelet coding 

Image Restoration and reconstruction

Image Restoration and Reconstruction

  1. Image enhancement is subjective and image restoration is objective.
  2. Goal is to improve the image in some predefined sense.
  3. Attempt to recover a degraded image by using  the knowledge of the degradation phenomenon which is usually available before hand.
  4. Understanding the degradation process and then modeling it is the key to success here.
  5. Example : Given y[m, n] and we are able to model y[m, n] = x[m, n] +η(m, n) where η = noise. We can get x[m, n] if we are able to understand η which is noise and which had degraded the original x[m, n]. By performing the inverse process of degradation on  y[m, n]  the original    x[m, n] can be restored.
  6. In reality, one can only estimate η. The quality of restoration largely depends on the closeness of the estimate to η.
  7. As in image enhancement, image restoration techniques are best formulated in the spatial domain, while others are better suited for the frequency domain.
  8. Rule of thumb : If the noise is additive use spatial domain technique and if the degradation is a motion induced blur,  use frequency domain techniques.
  9. However, additive noise is also taken care of in the frequency domain.
  10. Image restoration is divided into subtopics as follows :
a.       A model of the image degradation/restoration process
b.       Understanding how to model noise
c.       Understand the PDF of some important noise distribution such as Gaussian, Rayleigh, Erlang(gamma), Exponential, Uniform, Impulse(salt-and-pepper)  noise and finally periodic noise.
  1. The process of image restoration is given below

  1. Noise gets added to an image at the time of image acquisition and/or transmission.
  2. Factors include environmental conditions, quality of sensing elements, light levels, sensor temperature during acquisition.
  3. Interference in the channel, corruption due to lightning or other atmospheric disturbance during transmission.
  4. Noise models : Can be undertood through the corresponding Probability Density Function (PDF). Some of the most common and important noise PDFs are for the following : Gaussian noise, Rayleigh noise, Erlang(gamma) noise, Exponential noise, Uniform noise, Impulse (salt-and-pepper noise).
  5. The above PDFs, as a group, provide useful tools for modeling a broad range of noise corruption  situations found in practice.
  6. Gaussian noise is due to electronic circuit noise and sensor noise due to poor illumination and/or high temperature.
  7. Rayleigh noise  density is helpful in range imaging.
  8. Exponential and gamma densities find application in laser imaging.
  9. Impulse  noise is found in situations where quick transients, such as faulty switching, take place during imaging.
  10. An important observation is that it is difficult to differentiate visually between the first five noisy images even though their histograms are significantly different.
  11. The salt-and-pepper appearance of the image corrupted by impulse noise is the only one that is visually indicative of the type of noise causing the degradation.
  12. Periodic noise in an image arises typically from electrical or electromechanical interference during image acquisition. This is a spatially dependent noise which can be reduced significantly via frequency domain filtering.
  13. Restoration in the presence of noise is possible through spatial filtering.
  14. Spatial filters are of three types
a.      Mean filters :  Arithmetic Mean filter, Geometric Man Filter, Harmonic Mean Filter, Contraharmonic mean Filter
b.      Order-Statistic Filters : Median Filter, Max filter, Min filter, Midpoint Filter, Alpha-trimmed mean filter.
c.       Adaptive Filters : Adaptive local noise reduction filter,  Adaptive Median Filter
  1. Periodic noise reduction using Frequency domain filtering : These include Bandreject filters, Bandpass filters, Notch filters, Optimum notch filters.
  2. Estimating the Degradation function is done as follows :
a.       Estimation  by image observation
b.       Estimation by Experimentation.
c.       Estimation by modeling
  1. Inverse Filtering in general has poor performance and is improved by the following three methods :
a.       Minimum Mean Square Error (Wiener) Filtering makes provision for the degradation function and statistical characteristics of noise into the restoration process. Here, images and noise are considered as random variables and the objective is to minimise the mean square error
b.       Constrained Least Squares Filtering
c.       Geometric Mean Filter

Thursday, August 11, 2011

Digital Image Processing : Sample Practise Questions

The following questions have numbering for each section. Ignore the numbering and try to find answers for the given questions.

1. Draw the cross section of the human eye.
2. Explain- photopic or bright-light vision.
3. Explain-scotopic or dim-light vision.
4. What is called blind spot in the eye?
5. The observer is looking at a tree 15m high at a distance of 100m. Find the size of the retinal image.
6. Define Weber ratio.
7. Explain image formation in the eye.
8. What is brightness adaptation?
9. What is brightness discrimination?
10. What is Mach band pattern?
11. Explain simultaneous contrast.
12. How a digital image is represented?
13. What do you mean by a pixel?
14. What are called 4-neighbors of a pixel?
15. What are diagonal neighbors of a pixel?
16. Define 4-connectivity.
17. Define 8-connectivity.
18. Define m-connectivity.
19. What is called Euclidean distance?
20. What is called D4 distance or city-block distance? 
21. What is called D8 distance or chess board distance?
22. Give any four arithmetic operators with pixels.
23. Give any four logical operators with pixels.
24. What is meant by perspective transformation?
25. What is called camera calibration?.
26. What do you mean by stereo imaging?
27. What is called gray scale?

1. Explain in detail the elements of Digital Image Processing systems.
2. Explain in detail the structure of the human eye.
3. Explain image formation in the eye, brightness adaptation and discrimination.
4. Explain any four basic relationships between pixels.
5. Using perspective transformation and assuming the world coordinate system is aligned with the camera coordinate system, derive expression for x,y,z.
6. Using perspective transformation and assuming the world coordinate system is aligned with the camera coordinate system, derive expression for X,Y,Z.
7. Explain the basic transformations: translation, scaling, rotation, concatenation and inverse transformation.

1. Define a two dimensional bandlimited image.
2. Define Nyquist rate.
3. What is aliasing? Explain it with image spectrum.
4. What are called fold over frequencies?
5. Define sampling theorem.
6. List the practical limitations in sampling and reconstruction.
7. What is Moiré effect?
8. Write a note on display aperture.
9. What is quantization? How it is applied in images?
10. List the properties of the optimum mean square Quantizer.
11. What will be the mean square error for uniform PDF Optimal Quantizer?
12. What is meant by visual quantization?
13. How to reduce the contouring effect in visual quantization?
14. Draw the block diagram of Contrast quantization.
15. Draw the block diagram of pseudo random noise quantization.
16. What is a compandor?

1. Explain 2-D sampling theory in detail.
2. Explain in detail the practical limitations in sampling and reconstruction.
3. Explain the optimum mean square or Lloyd-max quantizer in detail.
4. Explain in detail about visual quantization.
5. Explain the process of reconstruction of an image from its samples.
6. Explain compandor design.

1) Properties of Fourier Transform?
2) Write the equation for 2D Fourier transform and Inverse Fourier transform?
3) Write the equation for the 2-D convolution?
4) Properties of convolution?
5) Define orthogonal matrix and unitary matrix?

7) If A and B are M1 xM2 and N1 xN2 matrices respectively, find their Kronecker product.
8) Define Kronecker Product?
10) Define the mean and co-variance of random signal?
11) For a given orthogonal matrix A and image U, 
Find the transformed image and the inverse transformation.
12) Write the equation for the 2-D orthogonal and unitary transforms?
13) Give some properties of 2-D DFT?
14) Write the equation for 2-D IDFT?
15) Mention the properties of Discrete cosine transform.
16) Define DCT?
17) Define Sequency?
18) Properties of Hadamard Transform?
19) For the 2x2 transform A and the image U

Calculate the transformed image V and the basis images.
20) What is Walsh transform?
21) Give some properties of 2-dimensional DFT?
22) What are the properties of walsh transform?
23) What is Haar Transform?
24) What is Hadamard transform?
25) Write the properties of Haar transform?
26) What is slant transform?
27) What are the properties of slant transform?
28) What is KL transform?
29) What are the properties of KL transform?
30) Compare KL, Hadamard, Haar and slant transform?


1) Find the Fourier spectrum , phase angle and power spectrum of 2D Fourier transform.
2) Write the DFT for the 2-variable? Find the DFT for the following sequence {2,3,4,4} and find its fourier spectrum?
3) Explain the following DFT properties:
a) Separability b) Translation
c) Rotation d) Distributivity & scaling
4) Explain the following (a) Convolution (b) Correlation
5) Develop an FFT algorithm for a 2-D transform? How many additions & multiplication are needed to compute 2D FFT of an N x N image?
6) Obtain the Haar transform matrix for N=4?
7) Obtain the Haar transform matrix for N=8?
8) Obtain the slant transform matrix for N=4?
9) Obtain the slant transform matrix for N=8?
10) Obtain the Hadamard transform matrix for N=4?
11) Obtain the Hadamard transform for N=8?
12) Obtain the Walsh transform matrix for N=4?


Part- A
1. What is image enhancement? What are the types of enhancement available?
2. What is point processing?
3. What is contrast stretching? Why is it required?
4. What is the need for reducing the dynamic range of image?
5. What is gray level slicing?
6. What is bit plane slicing?
7. What is histogram?
8. What is histogram equalization?
9. Draw the histogram for bright image and low contrast image?
10. Give the steps in histogram specification?
11. What is spatial filtering?
12. What are the effects of low pass and high pass filtering of images?
13. What is image smoothening? What are the types available?
14. Explain high boost filtering.
15. Draw the mask for different derivative filters.
16. What is homomorphic filtering?
17. Differentiate restoration and enhancement.
18. What is circulant and block-circulant matrices.
19. What is meant by diagonalization? Why is it required?
20. What is meant by algebraic approach to restoration? What are the types?
21. What is constrained and unconstrained restoration?
22. What is inverse filtering?
23. Write the equation for wiener filter.
24. What is parametric wiener filter?
25. What is rubber sheet transformation?
26. What are the basic operations in geometric transformation?
27. What is Gray- level interpolation?

Part –B
1) Explain the different point processing techniques used for image enhancement.
2) With neat diagram histogram processing and Equalization.
3) Explain histogram specification.
4) Explain the various types of image smoothening filters.
5) Explain in detail about homomorphic filtering.
6) Explain the concept of degradation model.
7) Explain discrete formulation of degradation model.
8) Describe the process involved in diagonalisation of circulant and block circulant matrices.
9) Explain constrained and unconstrained algebraic approach in detail.
10) Explain inverse filtering in detail.
11) Explain weiner filter in detail.
12) Write notes on Geometric transformation.