Presentation is loading. Please wait.

Presentation is loading. Please wait.

Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Saturday, Oct. 12, 2001 Haipeng.

Similar presentations


Presentation on theme: "Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Saturday, Oct. 12, 2001 Haipeng."— Presentation transcript:

1 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Saturday, Oct. 12, 2001 Haipeng Guo KDD Research Group Department of Computing and Information Sciences Kansas State University Fractal and Bayesian Networks Inference KDD Group Presentation

2 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Presentation Outline Simple Tutorial on Fractals Bayesian Networks Inference Review Joint Probability Space’s Fractal Property and its Possible Application to BBN Inference Summary

3 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Part I Introduction to Fractals

4 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractals Introduction Definition Examples: Man-made & Nature Fractals Fractal Dimension Fractal Applications

5 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference “ I coined fractal from the Latin adjective fractus. The corresponding Latin verb frangere means "to break" to create irregular fragments. It is therefore sensible - and how appropriate for our need ! - that, in addition to "fragmented" (as in fraction or refraction), fractus should also mean "irregular", both meanings being preserved in fragment. ” Fractal – “broken, fragmented, irregular” B. Mandelbrot : The fractal Geometry of Nature, 1982

6 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Definition: Self-similarity –A geometric shape that has the property of self-similarity, that is, each part of the shape is a smaller version of the whole shape. Examples:

7 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Mathematical fractal: Konch Snowflake Step One. Start with a large equilateral triangle. Step Two. Make a Star. 1.Divide one side of the triangle into three parts and remove the middle section. 2. Replace it with two lines the same length as the section you removed. 3. Do this to all three sides of the triangle. Repeat this process infinitely. The snowflake has a finite area bounded by a perimeter of infinite length !

8 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Real world fractals A cloud, a mountain, a flower, a tree or a coastline… The coastline of Britain

9 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractal geometry: the language of nature Euclid geometry: cold and dry Nature: complex, irregular, fragmented “Clouds are not spheres, mountains are not cones, coastlines are not circles, and bark is not smooth, nor does lightning travel in a straight line.”

10 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Euclid dimension In Euclid geometry, dimensions of objects are defined by integer numbers. 0 - A point 1 - A curve or line 2 - Triangles, circles or surfaces 3 - Spheres, cubes and other solids

11 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractal dimension Fractal dimension can be non-integers Intuitively, we can represent the fractal dimension as a measure of how much space the fractal occupies. Given a curve, we can transform it into 'n' parts (n actually represents the number of segments), and the whole being 's' times the length of each of the parts. The fractal dimension is then : d = log n / log s

12 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Example: Knoch snowflake After the first step, we get four segments(it's then divided into 4 parts). The whole curve is composed of three of these new segments. So, the fractal dimension is : d=log 4/log 3=1.2618... It takes more space than a 1 dimensional line segment, but it occupies less space than a filled two-dimensional square.

13 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Another example: Cantor Set The oldest, simplest, most famous fractal 1 We begin with the closed interval [0,1]. 2 Now we remove the open interval (1/3,2/3); leaving two closed intervals behind. 3 We repeat the procedure, removing the "open middle third" of each of these intervals 4 And continue infinitely. Fractal dimension: D = log 2 / log 3 = 0.63… Uncountable points, zero length Challenge problem: is ¾ in Cantor set?

14 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Devil’s staircase Take a Cantor Set, which is composed of an infinite number of points. Consider turning those points into dots and letting a Pacman eat them As our Pacman eats the dots, he gets heavier. Imagine that his weight after eats all the dots is 1. Graph his weight with time, we get devil’s staircase.

15 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Cantor square Fractal dimension: d = log 4 / log 3 = 1.26

16 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The Mandelbrot Set The Mandelbrot set is a connected set of points in the complex plane Calculate: Z 1 = Z 0 2 + Z 0, Z 2 = Z 1 2 + Z 0, Z 3 = Z 2 2 + Z 0 If the sequence Z 0, Z 1, Z 2, Z 3,... remains within a distance of 2 of the origin forever, then the point Z 0 is said to be in the Mandelbrot set. If the sequence diverges from the origin, then the point is not in the set

17 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Colored Mandelbrot Set The colors are added to the points that are not inside the set. Then we just zoom in on it

18 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Applications of fractals Astronomy: the struture of our universe Superclusters - clusters – galaxies- star systems(Solar system) - planets - moons Every detail of the universe shows the same clustering patterns. It can be modeled by random Cantor square The fractal dimension of our universe: 1.23

19 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Applications of fractals Rings of Saturn Originally, it was believed that the ring is a single one. After some time, a break in the middle was discovered, and scientists considered it to have 2 rings. However, when Voyager I approached Saturn, it discovered that the two ring were also broken in the middle, and the 4 smaller rings were broken as well. Eventually, it identified a very large number of breaks, which continuously broke even small rings into smaller pieces. The overall structure is amazingly similar to... Cantor Set

20 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Application of Fractals Human Body THE LUNGS: Formed by splitting lines Fractal Canopies The brain : The surface of the brain contains a large number of folds Human, the most intelligent animal, has the most folded surface of the brain Geometrically, the increase in folding means the increase in dimension In humans, it is obviously the highest, being as large as between 2.73 - 2.79

21 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Plants a tree branch looks similar to the entire tree a fern leaf looks almost identical to the entire fern One classic way of creating fractal plants is by means of l- systems(Lindenmayer)

22 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Bacteria Cultures A bacteria culture is all bacteria that originated from a single ancestor and are living in the same place. When a culture is growing, it spreads outwards in different directions from the place where the original organism was placed. The spreading of bacteria can be modeled by fractals such as the diffusion fractals

23 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Data Compression A color full-screen GIF image of Mandelbrot Set occupies about 35 kilobytes Formula z = z^2 + c, 7 bytes! (99.98% ) It could work for any other photos as well The goal is too find functions, each of which produces some part of the image. IFS (Iterated function system) is the key.

24 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Weather Weather behaves very unpredictably Sometimes, it changes very smoothly. Other times, however, it changes very rapidly Edward Lorenz came up with three formulas that could model the changes of the weather. These formulas are used to create a 3D strange attractor, they form the famous Lorenz Attractor, which is a fractal pattern.

25 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractal Antenna Practical shrinkage of 2-4 times are realizable for acceptable performance. Smaller, but even better performance

26 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Electronic Transmission Error During electronic transmissions, electronic noise would sometimes interfere with the transmitted data. Although making the signal more powerful would drown out some of this harmful noise, some of it persisted, creating errors during transmissions. Errors occurred in clusters; an period of no errors would be followed by a period with many errors. On any scale of magnification(month, day, hour, 20 minutes, …), the proportion of error- free transmission to error-ridden transmission stays constant. Mandelbrot studied the mathematical process that enables us to create random Cantor dust describing perfectly well the fractal structure of the batches of errors on computer lines

27 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Network Traffic Model Packets delays gain as a function of time in a WAN environment: the top diagram - absolute values of RTT parameter in virtual channel; the bottom diagram - fractal structure of packets flow that excessed 600 msec threshold.

28 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractal Art

29 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Fractal Summary Fractals are self-similar or self-affine structures Fractal object has a fractal dimension It models many natural objects and processes. It is the nature’s language. It has very broad applications.

30 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Part II Bayesian Networks

31 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Bayesian Networks Review Bayesian Networks Examples Belief Update and Belief Revision The joint Probability Space and Brute Force Inference

32 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Bayesian Networks Bayesian Networks, also called Bayesian Belief networks, causal networks, or probabilistic networks, are a network-based framework for representing and analyzing causal models involving uncertainty A BBN is a directed acyclic graph (DAG) with conditional probabilities for each node. –Nodes represent random variables in a problem domain –Arcs represent conditional dependence relationship among these variables. –Each node contains a CPT(Conditional Probabilistic Table) that contains probabilities of this node being specific values given the values of its parent nodes.

33 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Family-Out Example " Suppose when I go home at night, I want to know if my family is home before I try the doors.(Perhaps the most convenient door to enter is double locked when nobody is home.) Now, often when my wife leaves the houses, she turns on an outdoor light. However, she sometimes turns on the lights if she is expecting a guest. Also, we have a dog. When nobody is home, the dog is put in the back yard. The same is true if the dog has bowel problems. Finally, if the dog is in the back yard, I will probably hear her barking(or what I think is her barking), but sometimes I can be confused by other dogs. " Family-Out Light-On Bowel-problem Dog-out Hear-bark

34 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Asia Example from Medical Diagnostics

35 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Why is BBN important? Offers a compact, intuitive, and efficient graphical representation of dependence relations between entities of a problem domain. (model the world in a more natural way than Rule-based systems and neural network) Handle uncertainty knowledge in mathematically rigorous yet efficient and simple way Provides a computational architecture for computing the impact of evidence nodes on beliefs(probabilities) of interested query nodes Growing numbers of creative applications

36 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Alarm Example: the power of BBN PCWP CO HRBP HREKG HRSAT ERRCAUTER HR HISTORY CATECHOL SAO2 EXPCO2 ARTCO2 VENTALV VENTLUNG VENITUBE DISCONNECT MINVOLSET VENTMACH KINKEDTUBE INTUBATIONPULMEMBOLUS PAPSHUNT ANAPHYLAXIS MINOVL PVSAT FIO2 PRESS INSUFFANESTHTPR LVFAILURE ERRBLOWOUTPUT STROEVOLUMELVEDVOLUME HYPOVOLEMIA CVP BP The Alarm network 37 variables, 509 parameters (instead of 2 37 )

37 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Applications Medical diagnostic systems Real-time weapons scheduling Jet-engines fault diagnosis Intel processor fault diagnosis (Intel); Generator monitoring expert system (General Electric); Software troubleshooting (Microsoft office assistant, Win98 print troubleshooting) Space shuttle engines monitoring(Vista project) Biological sequences analysis and classification ……

38 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Bayesian Networks Inference Given an observed evidence, do some computation to answer queries An evidence e is an assignment of values to a set of variables E in the domain, E = { X k+1, …, X n } –For example, E = e : { Visit Asia = True, Smoke = True} Queries: –The posteriori belief: compute the conditional probability of a variable given the evidence, P(Lung Cancer| Visit Asia = TRUE AND Smoke = TRUE) = ?  This kind of inference tasks is called Belief Updating –MPE: compute the Most Probable Explanation given the evidence  An explanation for the evidence is a complete assignment { X 1 = x 1, …, X n = x n } that is consistent with evidence. Computing a MPE is finding an explanation such that no other explanation has higher probability  This kind of inference tasks is called Belief revision

39 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Belief Updating The problem is to compute P(X=x|E=e): the probability of query nodes X, given the observed value of evidence nodes E = e. For example: Suppose that a patient arrives and it is known for certain that he has recently visited Asia and has dyspnea. - What’s the impact that this evidence has on the probabilities of the other variables in the network ? P(Lung Cancer) = ? Visit to Asia Smoking Lung Cancer Tuberculosis tub. or lung cancer Bronchitis X-Ray Dyspnea

40 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Belief Revision Let W is the set of all nodes in our given Bayesian network Let the evidence e be the observation that the roses are okay. Our goal is to now determine the assignment to all nodes which maximizes P(w|e). We only need to consider assignments where the node roses is set to okay and maximize P(w), i.e. the most likely “state of the world” given the evidence that rose is ok in “this world”. The best solution then becomes - P(sprinklers = F, rain = T, street = wet, lawn = wet, soil = wet, roses = okay) = 0.2646

41 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Complexity of BBN Inference Probabilistic Inference Using Belief Networks is NP-hard. [Cooper 1990] Approximating Probabilistic Inference in Bayesian Belief Networks is NP-hard [Dagum 1993] Hardness does not mean we cannot solve inference. It implies that –We cannot find a general procedure that works efficiently for all networks –However, for particular families of networks, we can have provably efficient algorithms either exact or approximate –Instead of a general exact algorithm, we seek for special case, average case, approximate algorithms –Various of approximate, heuristic, hybrid and special case algorithms should be taken into consideration

42 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference BBN Inference Algorithms Exact algorithms –Pearl’s message propagation algorithm(for single connected networks only) –Variable elimination –Cutset conditioning –Clique tree clustering –SPI(Symbolic Probabilistic Inference) Approximate algorithms –Partial evaluation methods by performing exact inference partially –Variational approach by exploiting averaging phenomena in dense networks(law of large numbers) –Search based algorithms by converting inference problem to an optimization problem, then using heuristic search to solve it –Stochastic sampling also called Monte Carlo algorithms

43 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Inference Algorithm Conclusions The general problem of exact inference is NP-Hard. The general problem of approximate inference is NP-Hard. Exact inference works for small, sparse networks only. No single champion either exact or inference algorithms. The goal of research should be that of identifying effective approximate techniques that work well in large classes of problems. Another direction is the integration of various kinds of approximate and exact algorithms exploiting the best characteristics of each algorithm.

44 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Part III BBN Inference using fractal? Joint Probability Distributions Space of a BBN Asymmetries in Joint Probability Distributions (JPD) Fractal Property of JPD How does it help to do approximate inference?

45 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Joint Probability Distribution Asia revisited: - Eight binary nodes - Each node has 2 states: Y or N - Total states: 2 8 = 256 Instance# a b c d e f g h Probability Instance 0: 0 0 0 0 0 0 0 0 0.000274 Instance 1 : 0 0 0 0 0 0 0 1 0.000118 Instance 2 : 0 0 0 0 0 0 1 0 0 Instance 3 : 0 0 0 0 0 0 1 1 0 Instance 4 : 0 0 0 0 0 1 0 0 0 Instance 5 : 0 0 0 0 0 1 0 1 0 Instance 6 : 0 0 0 0 0 1 1 0 0 Instance 7 : 0 0 0 0 0 1 1 1 0 ……….. Instance 252 : 1 1 1 1 1 1 0 0 0.001698023 Instance 253 : 1 1 1 1 1 1 0 1 0.015282209 Instance 254 : 1 1 1 1 1 1 1 0 0.032262442 Instance 255 : 1 1 1 1 1 1 1 1 0.290361976 Visit to Asia Smoking Lung Cancer Tuberculosis tub. or lung cancer Bronchitis X-Ray Dyspnea

46 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference JPD of Asia 256 states Max: 0.290362 Spreads over 9 orders of magnitude 0,1.50E-09, …, 0.290362 Top 30 most likely states cover 99.2524% of the total probability space. Conclusion: There is usually a small set fraction of states that covers a large portion of the total Probability space with the remaining states having Practically negligible probabilities.

47 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Why is the JPD so skew? When we know nothing about the domain, JPD should be flat. The more we know about a domain, the more asymmetry individual probabilities will show. When the domain and its mechanism are well-known, probability distributions tend to be extreme. Conclusion: Asymmetries in the individual distributions result in joint probability distributions exhibiting orders of magnitude differences in probabilities of various states of the model.

48 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference How does this help inference? Considering only a small number of probabilities of individual states can lead to good approximations in Belief updating. The result can be refined by exploring more high likely states. Problem: Where to locate these “peaks”?

49 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The global map of JPD of Asia To locate these peaks, let’s first make the map. Nodes order: visit to Asia?|smoking?|tuberculosis?|either tub. or lung cancer?|positive X-ray?|lung cancer?|bronchitis?|dyspnoea? CPT arrangements: let small numbers go first, for example 0.01 0.99, 0.3 0.7 in order to shift high values to the same area. Most “peaks” are in the second half of the map Clusters of high “peaks”

50 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Self Similarity (or self affine) Property

51 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Self Similarity (or self affine) Property Scale first half(256,0-128, 0-64, 0-32)

52 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Self Similarity (or self affine) Property Scale second half(1-256,129-256, 193-256, 224-256)

53 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference More Observations How the skewed “map” is being formed? First, suppose we know nothing about the domain, i.e., all numbers are 0.5 Then, we explore the nodes one by one to set the numbers of its CPT to its actual values (most likely asymmetry, for example, 0.01/0.99, 0.2/0.8). We draw the JPD after each node’s CPT is updated. Thus we get 8 graphs.

54 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The process of getting skew JPD

55 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Putting together

56 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference What fractal is this? Cantor Set Recall

57 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The Devil's Staircases

58 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The process of building Devil’s Staircase (Asia)

59 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Multifractal Consider again the Cantor set, but let the original bar have a uniformly distributed mass 1. Suppose the mass is conserved in the bisection process. Suppose at each bisection step, the mass carried by any element is assigned in an asymmetry wat, say, to the left with probability P = 0.25. 1 0.25 0.75 0.25x0.25, 0.75x0.25 0.25x0.75, 0.75x0.75 0.25x0.25x0.25 0.25x0.25x0.75 …… …… …… …… 0.25x0.75x0.75,0.75x0.75x0.75

60 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Random Multifractal? –Both are self affine –Both have clusters of high peaks –More mass distributions assigned to the right end –BBN JPD more irregular, more random, the probability distributions are more extreme.

61 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The cumulative Mass Distributions Three Devil’s Staircases:

62 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Future Problems Construct the proper “random multifractal models” for given Bayesian networks. Locate these “clusters of high peaks” from the fractal model. Once we know that information, other algorithms, for example GA, can be used to explore the local optimal. Identify “Backbone nodes” to help to find out these “high peaks” Use these “high peaks” to do approximate inference. Design an anytime inference algorithm from the above scheme Identify the characteristics of BBNs on which this algorithms work best(should be related to the asymmetry of these BBNs’ CPTs).

63 Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference The End Any Questions ? Thank you!


Download ppt "Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Saturday, Oct. 12, 2001 Haipeng."

Similar presentations


Ads by Google