Image processing in the compressed domain Assist.Eng. Camelia Florea, Technical University of Cluj-Napoca, ROMANIA SSIP 2009.

Slides:



Advertisements
Similar presentations
JPEG Compresses real images Standard set by the Joint Photographic Experts Group in 1991.
Advertisements

Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Pattern Recognition and Machine Learning
Digital Image Processing
Robust statistical method for background extraction in image segmentation Doug Keen March 29, 2001.
Computer vision: models, learning and inference Chapter 8 Regression.
Image Processing IB Paper 8 – Part A Ognjen Arandjelović Ognjen Arandjelović
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
SWE 423: Multimedia Systems
Segmentation Divide the image into segments. Each segment:
JPEG.
CS :: Fall 2003 MPEG-1 Video (Part 1) Ketan Mayer-Patel.
Image Enhancement.
CMPT 365 Multimedia Systems
T.Sharon-A.Frank 1 Multimedia Image Compression 2 T.Sharon-A.Frank Coding Techniques – Hybrid.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
5. 1 JPEG “ JPEG ” is Joint Photographic Experts Group. compresses pictures which don't have sharp changes e.g. landscape pictures. May lose some of the.
1 JPEG Compression CSC361/661 Burg/Wong. 2 Fact about JPEG Compression JPEG stands for Joint Photographic Experts Group JPEG compression is used with.jpg.
Image Compression - JPEG. Video Compression MPEG –Audio compression Lossy / perceptually lossless / lossless 3 layers Models based on speech generation.
Trevor McCasland Arch Kelley.  Goal: reduce the size of stored files and data while retaining all necessary perceptual information  Used to create an.
Lossy Compression Based on spatial redundancy Measure of spatial redundancy: 2D covariance Cov X (i,j)=  2 e -  (i*i+j*j) Vertical correlation   
CS559-Computer Graphics Copyright Stephen Chenney Image File Formats How big is the image? –All files in some way store width and height How is the image.
Entropy and some applications in image processing Neucimar J. Leite Institute of Computing
Introduction to JPEG Alireza Shafaei ( ) Fall 2005.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 8 – JPEG Compression (Part 3) Klara Nahrstedt Spring 2012.
Klara Nahrstedt Spring 2011
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
JPEG. The JPEG Standard JPEG is an image compression standard which was accepted as an international standard in  Developed by the Joint Photographic.
Indiana University Purdue University Fort Wayne Hongli Luo
JPEG CIS 658 Fall 2005.
Automatic Minirhizotron Root Image Analysis Using Two-Dimensional Matched Filtering and Local Entropy Thresholding Presented by Guang Zeng.
November 13, 2014Computer Vision Lecture 17: Object Recognition I 1 Today we will move on to… Object Recognition.
Eyes detection in compressed domain using classification Eng. Alexandru POPA Technical University of Cluj-Napoca Faculty.
Lecture 7: Intro to Computer Graphics. Remember…… DIGITAL - Digital means discrete. DIGITAL - Digital means discrete. Digital representation is comprised.
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
The task of compression consists of two components, an encoding algorithm that takes a file and generates a “compressed” representation (hopefully with.
Digital Image Processing
JPEG.
CS654: Digital Image Analysis
Adaptive Wavelet Packet Models for Texture Description and Segmentation. Karen Brady, Ian Jermyn, Josiane Zerubia Projet Ariana - INRIA/I3S/UNSA June 5,
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 11 COMPRESSION.
Wonjun Kim and Changick Kim, Member, IEEE
Chapter 8 Lossy Compression Algorithms. Fundamentals of Multimedia, Chapter Introduction Lossless compression algorithms do not deliver compression.
Introduction to JPEG m Akram Ben Ahmed
JPEG. Introduction JPEG (Joint Photographic Experts Group) Basic Concept Data compression is performed in the frequency domain. Low frequency components.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
By Dr. Hadi AL Saadi Lossy Compression. Source coding is based on changing of the original image content. Also called semantic-based coding High compression.
IS502:M ULTIMEDIA D ESIGN FOR I NFORMATION S YSTEM M ULTIMEDIA OF D ATA C OMPRESSION Presenter Name: Mahmood A.Moneim Supervised By: Prof. Hesham A.Hefny.
Image Processing Architecture, © Oleh TretiakPage 1Lecture 5 ECEC 453 Image Processing Architecture Lecture 5, 1/22/2004 Rate-Distortion Theory,
Chapter 8 Lossy Compression Algorithms
Biointelligence Laboratory, Seoul National University
JPEG Compression What is JPEG? Motivation
IMAGE COMPRESSION.
Chapter 9 Image Compression Standards
JPEG Image Coding Standard
Discrete Cosine Transform
JPEG.
CMPT 365 Multimedia Systems
Image Segmentation Techniques
JPEG Pasi Fränti
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
JPEG Still Image Data Compression Standard
Pattern Recognition and Machine Learning
Image Coding and Compression
Generally Discriminant Analysis
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Image processing in the compressed domain Assist.Eng. Camelia Florea, Technical University of Cluj-Napoca, ROMANIA SSIP 2009

Contents Brief Overview JPEG standard coding The two ways to process JPEG compressed images DCT Coefficients JPEG features space for image segmentation A DCT-based approach for detecting patients‘ identification information Bayesian segmentation of hepatic biopsy color images in the JPEG compressed domain Image enhancement operations in the compressed domain Compressed domain implementation of fuzzy rule-based contrast enhancement

Brief Overview Compressed domain image processing algorithms in encoded JPEG image domain - provide a powerful computational alternative to classical. This field is in its beginning - the algorithms reported in the literature are mostly based on linear arithmetic point operations (addition, substraction, multiplication). Advantage: no need to decompress/ recompress the whole image prior to processing/after processing. we compute less data (after quantization many of the DCT coefficients are zero).

JPEG standard coding The color space of the image is converted (RGB to YUV) The image is divided into 8x8 blocks Values are scaled symmetrical towards 0 (from [0, 255] to [-128, 127]) Each 8x8 block is processed for compression DCT is applied on each block => obtain the DCT coefficients (DC and AC) DCT coefficients are quantized - small coefficients are quantized to zero zig-zag scan of the DCT blocks RLE (Run Length Encoding) is performed finally - entropy coding

The image in the JPEG compressed domain

The two ways to process JPEG compressed images Having a JPEG compressed image is more efficiently to process it, without: performing decompression, pixel level processing, and recompression. Processing in the compressed domain is made over the RLE vectors. RLE vector contains data about the variation and mean of luminance/color

DCT Coefficients There are many local texture features embedded in the DCT coefficients, reflecting color and texture: the DC coefficient - the average color in a block of pixels, the AC coefficients - the variance of luminance and chrominance Two dimensional DCT basis functions (N = 8).

JPEG features space for image segmentation Image segmentation is the technique of partitioning an image into units which are homogeneous with respect to one or more characteristics. This can be done on pixel level for highest accuracy, but also on JPEG block level, in the compressed domain, considering the local color and texture information roughly needed, is completely present in this representation.

A DCT-based approach for detecting patients‘ identification information Addressed problem: implementing an algorithm for detecting patient’s data from JPEG ultrasound images, applied directly on DCT coefficients. Advantage: no need to convert medical images/video back to the spatial (uncompressed) domain. the algorithm can detect textual information using the amount of energy, computed using only AC coefficients. HIPAA recommends to healthcare providers: the protection of the confidentiality of their patients’ health data. Medical information, regarding both: patients’ identification and their treatment, can be: transmitted and stored and, are susceptible of being accessed by unauthorized people. Images contains textual data about the patient, data that requires special security measures when disseminating the images. They must be handled by authorized health care professionals only.

It is possible to hide or eliminate the patient information without processing the image content itself (pixels) – by using only the DCT coefficients. The DCT - is one of the best filters for feature extraction in the frequency domain – it could be used here. In an ultrasound medical image the areas with very high energy amount – are the regions containing textual information. Areas with patient’s data can then be encrypted, blurred or eliminated. Having a basic knowledge of the ultrasound machine used, + With the same image acquisition conditions, => we can detects (and hide) only the patient identification information in the image and keep the medical information. A DCT-based approach for detecting patients’ identification information

System’s architecture

Analyzing the grey level variation The block’s energy is computed and analyzed – for each 8×8 block. (E AC - the average of AC coeff. Energy) In ultrasound images are many 8×8 blocks where the local variation of the brightness is small: the background area, and the examination area from ultrasound image. => every such block - do not contain text information and, under no circumstance, information about patients. If the 8×8 blocks exhibit a large variation of the grey levels around the average brightness value, => we have areas with sudden changes of brightness from black to white: text, or cartesian axes, from the ultrasound image.

The data hiding algorithm in JPEG compressed domain 1.Compute the E AC. 2.If E AC data from the original image are kept (no processing). 3.If E AC ≥ e thd => the block has a significant content of details,  areas of interest - need to be processed for data selection: -patient’s identification information - will be protected, -examination data - no processing. -where e thd represents the optimal selection threshold between: -the uniform blocks, and -the blocks with a significant number of details.

High energy blocks Energy blocks Selected data to hide. Medical image resulted just by keeping only the blocks with high energy. High energy blocks Apply selection rules Post Processing

Bayesian segmentation of hepatic biopsy color images in the JPEG compressed domain Addressed problem color image segmentation based on: the color information, and, also on the local texture information, for each 8×8 pixel neighborhoods. Advantage: no need to decompress/recompress the whole image prior to processing/after processing. we compute less data (after quantization many of the DCT coefficients are zero). This reduced dimensionally feature space makes easier the training and implementation of rather complex classifiers, as e.g. the Bayesian classifier with class probabilities modeled by Gaussian mixtures used here.

Color in RGB vs. YUV space Many image representation spaces can be used in segmentation process. The YUV representation yields certain advantages over RGB: The YUV is the representation used in the JPEG standard, The YUV provides a clear separation between the luminance representation (Y) and color (U,V), The luminance information Y, the color information U and V exhibits poor correlation The image storage format itself provides the information needed for an accurate identification/segmentation e.g.: segmentation of the hepatic biopsies into tissue vs. microscopic slide and further, of the tissue into healthy tissue vs. hepatic fibrosis.

Bayesian classification of pixel block in the discrete cosine transform domain We use the Bayes decision rule to classify a DCT block into microscopic slide, healthy tissue or fibrosis (features space = zig-zag scanned quantized DCT coefficients). A powerful yet simple model for blocks classification is the Multivariate Gaussian model: where: The Bayes decision rules for minimal cost are the following: Typically in the hepatic biopsy there is no obvious reason to assume uneven distribution of the tissue vs. microscopic slide, and neither of fibrosis vs. healthy tissue, therefore, we consider:

Gaussian parameters estimation For the classification we need a-priori knowledge of the class statistics: If is square and singular, then its inverse does not exist. → This might be the case when the matrix is sparse, as is the case when using DCT quantized coefficients. In these cases, the computation is based on computing singular value decomposition of (the base of Moore-Penrose pseudo-inverse). Any singular values less than a threshold-value are treated as zero (>0.01). The determinant of matrix is the product of the diagonal singular values.

Bayesian segmentation of hepatic biopsy color images in the JPEG compressed domain The training phase of the algorithm The statistical properties of the classes used by the two classifiers are determined using ground-truth images: As a result the mean values and the covariance matrices (as well as their pseudo- inverse) are found for each class. The test phase of the algorithm Each and every 8×8 block from the microscopic compressed image is considered and the blocks are processed for classification. The segmentation of an image is performed as a 2-step classification process: 1 st step - Discrimination between microscopic slide and hepatic tissue 2 nd step - Identify the blocks that exhibit fibrosis among the hepatic tissue blocks

Discrimination between microscopic slide and hepatic tissue (1 st step) 1 st step is the discrimination between microscopic slide and hepatic tissue (with or without fibrosis). In this case the luminance information gives sufficient information for segmentation, and the decision rule is: with Y dct [8×8] – the matrix of the DCT coefficients

Identify the blocks that exhibit fibrosis among the hepatic tissue blocks (2 nd step) The hepatic biopsies are treated with Sirius stain → the coloration of hepatic fibrosis appears reddish – unlike the healthy tissue, of beige color. This 2-class Bayesian classification is performed at block level, but this time, the Y and V components are used to compute the two class probabilities U is not used since it is a measure of the dominance of blue Color components and luminance information are not correlated → the joint class probabilities are the product of the luminance and color probabilities: where: Y dct - the 8×8 block of the luminance DCT coefficients; V dct - the 8×8 block of the reddish chrominance DCT coefficients.

Experimental results Patient DCT algorithm [%] Pixel level algorithm [%] Scores of fibrosis P P P P P P P Classifications results using our algorithm and pixel level algorithm:

Experimental results 1 st classifier2 nd classifier FARFRRFARFRR Average Worse case FAR Worse case FRR False acceptance rate (FAR) and false rejection rate (FRR), for one patient:

Image enhancement operations in the compressed domain Mostly linear algorithms developed for the compressed domain: Pointwise image addition/substraction Constant addition/substraction to each spatial position Constant multiplication to each spatial position Pointwise image multiplication Pixel arithmetic can be used to implement a number of operations cross-fade between two images or video sequences image composition - overlaying a forecaster on a weather map implementation of fuzzy rule-based contrast enhancement

Alpha-blending between two images

Image composition

Compressed domain implementation of fuzzy rule-based contrast enhancement implementing a non-linear operator using compressed domain processing – fuzzy rule-based contrast enhancement,Takagi-Sugeno Advantage: no need to decompress/ recompress the whole image prior to processing/after processing. for the 8×8 size blocks processed in the compressed domain, the processing implies a single comparison of the coefficient with the threshold (instead of 64 comparisons needed at pixel level).

Description of the fuzzy rule-based contrast enhancement algorithm The fuzzy rule base of the Takagi-Sugeno fuzzy systems comprises the following 3 rules: R1: IF l u is Dark THEN l v is Darker R1: IF l u is Dark THEN l v =l v d R2: IF l u is Gray THEN l v is Midgray  R2: IF l u is Gray THEN l v =l v g R3: IF l u is Bright THEN l v is Brigter, R3: IF l u is Bright THEN l v =l v b Input and output membership functions for fuzzy rule-based contrast enhancement: For any value at the input of our Takagi-Sugeno contrast enhancement fuzzy system, in the output image, the corresponding brightness is obtained by applying the Takagi-Sugeno fuzzy inference, as: Where:,, denote the membership degrees of the currently processed brightness to the input fuzzy sets Dark, Gray and Bright.

The adaptive algorithm for contrast enhancement To obtain in the compressed domain the same processing results as the one given by the pixel-level approach - the algorithm must be reformulated as a block level processing. The nonlinear operations, like the thresholding in fuzzy rule-based contrast enhancement algorithm, must be carefully addressed. The DC coefficient gives the average brightness in the block - is used as an estimate for selecting the processing rule for all the pixels in the blocks with small AC energy. In this algorithm an adaptive minimal decompression is used: - full decompression is no longer needed, - but, decompression is used for the block having many details, for an improved accuracy of processing.

The fuzzy set parameters selection using the DC histogram in the compressed domain A reasonable choice for the thresholds values, and would be the minimum, the mean and the maximum grey level from the image histogram. Roughly speaking, if the DC coefficients would be the only ones used to reconstruct the pixel level representation (without any AC information), - they would give an approximation of the image, - with some block boundary effects and some loss of details, - but, however still preserving the significant visual information. Therefore, - the histogram built only from the DC coefficients will have approximately the same shape as the grey level histogram. Histogram of DC coefficients, and at pixel level (frog.jpg)

Experimental results Input membership function superimposed on the DC histogram of the Y component DC histogram of the Y component after fuzzy contrast enhancement The algorithm is applied only on the luminance component. However, it can be used to enhance color images as well, with no change of the chrominance components

Imagee thd EffBlocks [%] MSE frog.jpg woman.jpg Lena.jpg keyboard.jpg Results for different values e thd The Mean Squared Error (MSE ) between the pixel level processed images and the images processed with our algorithm was used as quality performance measure. The efficiency (EffBlocks) of the proposed method formulated above for the compressed domain, is evaluated by examining the number of blocks processed at pixel level as percent from the total number of 8×8 pixels blocks in the image.

SVM in the compressed domain