Presentation is loading. Please wait.

Presentation is loading. Please wait.

4.2 Data Input-Output Representation

Similar presentations


Presentation on theme: "4.2 Data Input-Output Representation"— Presentation transcript:

1 4.2 Data Input-Output Representation
Background ( Ref: Handbook of Neural Computation ) Saund: Key theme in AI is to discover good rep. for the problem at hand. A good rep. makes explicit info. useful to the computation, it strips away obscuring clutter, it reduces info. to its essentials. Raw Data Transf. NN Final Output Input Output Data Rep. is more important/critical than network topology Goals of Data Rep. [Garbage In Garbage Out] Feature Enhancement / Data Reduction for Separability Similar (Diff) events  Similar (Diff) rep. for better interpolation More elements for important features Compactness without losing (or even enhancing) info. for fast learning Preserve feature info. (clustering / metric info) Ex. Binary coding could destroy metric 01111  [Hamming metric]

2 Ex. Character Recognition
Raw data = 5 64bit binary vectors. Representation ① raw data ② any code for five characters ③ shape features : horizontal and vertical spars, their ratio, relative positions of the spars Other extreme: Wasserman : raw data may be more useful in cases where essential features are unknown. NN can discover features in hidden neurons.

3 2. Data Preprocessing techniques
Data sets are plagued by Noise, Bias, Large variations in dynamic range . . . (1) Normalize  Remove large dynamic variances over one or more dimensions in data Ex. ① Normalize gray scale image  invariant to light condition ② Normalize speech signal  invariant to absolute volume level ③ Normalize with respect to position and size - Character recognition

4 (3) Principal Component Analysis – Dim. Reduction
(2) Nomalization Algorithms : One Way to embed magnitude ( l ) info. :  Normalize ② Row Norm. (2-D) : for each row, divide by the mean value ③ Column Norm. (2-D) (3) Principal Component Analysis – Dim. Reduction

5 3. Case Study : Face Recognition
(profile) Data Reduction 416 x  x  23 (Discard High freq. components) (= 133,120 pixels) ① Efficient tech. to extract high interest features. ② Method for data reduction with minimal info. loss ③ DCT applied to reduced vector descriptions enhances info. content, provides invariance to small change and increased separability.

6 1 0 0 0 0 0 0 0 0 1 1 1 1 0 0 0 0 0 0 1 4. Actual Coding Schemes (1)
Local R O Y G B I V (2) Distributed C2 C1 C0 In Local representation, each node may be [0 1] [-1 1] or even continuous when more than one node can be active to indicate presence of two or more features. (3) Coarse Distributed Wider overlapping receptive fields 36 nodes nodes

7 Students’ Questions from 2005:
While DCT in facial image processing helps data reduction, does it also help face recognition ? Since Human faces are all alike, its Fourier transform will also be similar. Spatial features will be more relevant to recognition. Normalization or Coding will reduce data or help classification. But, isn’t the process going to delay the overall learning time ? Coarse distributed coding will reduce the total number of nodes. However, when a single node is represented in overlapped fashion, isn’t the additional info. needed such as the overlap position, etc. ?

8 When an NN technique is used for character or speech recognition, how does its performance compare with non-NN approaches ? NN can be applied to many problems. Any application where NN is hard to apply to ? Is there any general measure to tell the importance of information in feature extraction ? If line search is used to find an optimal learning rate, the number of steps may decrease but I am afraid the overall processing time may increase. Can better separation for classification result via data representation ? Can the information content increase via a good data rep. ?

9 5. Discrete Coding (1) Simple Sum (fault tolerance, but requires many nodes for large numbers) 5 = = = (2) Value Unit Encoding 2 = 10 = (3) Discrete Thermometer x >0 x>3 x>6 x>9 x>12 2 = 10 = x >0 x>3 x>6 x>9 x>12

10 6. Continuous Coding (1) Simple Analog
• For an activation ai range of [0,1] or [-1,1], value in range [u, v] = (v - u) ai + u • Logarithmic scale for data set with large dynamic range (2) Continuous Thermometer

11 Ref. “Neural Network Perception For Mobile Robot Guidance,”
(3) Proportional Coarse Coding Pomerleau used a Gaussian Smearing function to represent steering directions in an Autonomous Land Vehicle In a Neural Network (ALVINN) Ref. “Neural Network Perception For Mobile Robot Guidance,” D. Pomerleau, Kluwer, 93. Slight Right Turn


Download ppt "4.2 Data Input-Output Representation"

Similar presentations


Ads by Google