The Basics of Graphical Models


 Georgiana Crawford
 5 years ago
 Views:
Transcription
1 The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures are taken from this chapter. Consider a set of random variables fx 1 ; : : : ; X n g. We are interested in Which variables are independent? Which variables are conditionally independent given others? What are the marginal distributions of subsets of variables? These questions are answered with the joint distribution P.X 1 ; : : : ; X n /. Marginalization is answered by summing over the joint. Independence is answered by checking factorizations. Assume the variables are discrete (i.e., categorical). The joint distribution is a table p.x 1 ; : : : ; x n /. Each cell is nonnegative and the table sums to one. If there are r possible values for each variable then the naïve representation of this table contains r n elements. When n is large, this is expensive to store and to use. Graphical models provide a more economic representation of the joint distribution by taking advantage of local relationships between random variables. Directed graphical models A directed graphical model is a directed acyclic graph. The vertices are random variables X 1 ; : : : ; X n ; edges denote the parent of relationship, where i are the parents of X i. Here is an example: 1
2 X 4 X 2 X 1 X 6 X 3 X 5 The random variables are fx 1 ; : : : ; X 6 g, e.g., 6 D fx 5 ; X 2 g. The graph defines a factorization of the joint distribution in terms of the conditional distributions p.x i j x i /. In general, p.x 1W6 / p.x 1 /p.x 2 j x 1 /p.x 3 j x 1 /p.x 4 j x 2 /p.x 5 j x 3 /p.x 6 j x 2 ; x 5 / p.x 1Wn / ny p.x i j x i /: (1) id1 (Note that we can use a set in the subscript.) This joint is defined in terms of local probability tables. Each table contains the conditional probabilities of a variable for each value of the conditioning set. By filling in the specific values for the conditional distributions, we produce a specific joint distribution of the ensemble of random variables. Holding the graph fixed, we can change the local probability tables to obtain a different joint. Now consider all possible local probability tables. We see that the graphical model represents a family of distributions. The family is defined by those whose joint can be written in terms of the factorization implied by the graph. It is important to notice that this is not all distributions over the collection of random variables. Graphical models represent a family of distributions. What is the advantage of limiting the family? Suppose x 1Wn are binary random variables. The full joint requires 2 n values, one per entry. The graphical model joint requires P n id1 2j i j entries. We have replaced exponential growth in n by exponential growth in j i j. In statistical and machine learning applications, we represent data as random variables and analyze data via their joint distribution. We enjoy big savings when each data point only depends on a couple of parents. 2
3 This is only part of the story. In addition to economic representation: Graphical models give us inferential machinery for computing probabilistic quantities and answering questions about the joint, i.e., the graph. The graph determines, and thus lets us control, the cost of computation. (And, as an aside, these same considerations apply when thinking about data and statistical efficiency. But this is less looked at in the graphical models literature.) Finally, graphs are more generic than specific joint distributions. A graphical model of binary data can be treated with similar algorithms as a graphical model with rary data. And, later, we will see how the same algorithms can treat discrete / categorical variables similarly to continuous variables, two domains that were largely considered separately. For example, graphical models connect the algorithms used to work with hidden Markov models to those used to work with the Kalman filter. Directed graphical models and conditional independence Recall the definition of independence x A? x B! p.x A ; x B / D p.x A /p.x B / (2)! p.x A j x B / D p.x A / (3)! p.x B j x A / D p.x B / (4) And recall the equivalent definitions of conditional independence x A? x B j x C! p.x A ; x B j x C / D p.x A ; x B j x C / (5)! p.x A j x B ; x C / D p.x A j x C / (6)! p.x B j x A ; x C / D p.x B j x C / (7) These are questions about factorizations of marginal distributions. They can be answered by examining or computing about the graph. Recall the chain rule of probability p.x 1Wn / D ny p.x i j x 1 ; : : : ; x i 1 / (8) id1 In our example p.x 1W6 / D p.x 1 /p.x 2 j x 1 /p.x 3 j x 1 ; x 2 / p.x 6 j x 1W5 /: (9) The joint distribution defined by the graph is suggestive that, e.g., p.x 4 j x 1 ; x 2 ; x 3 / D p.x 4 j x 2 /; (10) 3
4 which means that x 4? x f1;3g j x 2 : (11) This statement is true. It is one of the basic conditional independence statements. Let s prove it: Write the conditional as a ratio of marginals p.x 4 j x 1 ; x 2 ; x 3 / D p.x 1; x 2 ; x 3 ; x 4 / p.x 1 ; x 2 ; x 3 / (12) Numerator: take the joint and marginalize out x 5 and x 6 Denominator: Further marginalize out x 4 from the result of the previous step. Finally, divide to show that this equals p.x 4 j x 2 /. More generally, let I be a topological ordering of the random variables, which ensures that i occurs in the ordering before i. Let i be the set of indices that appear before i, not including i. The set of basic conditional independence statements is fx i? x i j x i g (13) In our example, one valid topological ordering is I D f1; 2; 3; 4; 5; 6g. This implies the following independencies, X 1? 0 j 0 (14) X 2? 0 j X 1 (15) X 3? X 2 j X 1 (16) X 4? fx 1 ; X 3 g j X 2 (17) X 5? fx 1 ; X 2 ; X 4 g j X 3 (18) X 6? fx 1 ; X 3 ; X 4 g j fx 2 ; X 5 g (19) This is a little inelegant because it depends on an ordering. Here is a graphspecific definition of the basic conditional independencies. Redefine i to be all the ancestors of i excluding the parents. Using this definition, we can define the basic conditional independencies as in Equation (13). Note: This does not give us all of the possible basic conditional independencies, i.e., all of those possible by traversing all topological orderings, but that doesn t really matter.) The point is this. We can read off conditional independencies by looking at the graph. We emphasize that these independencies hold regardless of the specific local probability tables. This gives us an insight about the nature of graphical models: 4
5 By using the cheaper factorized representation of the joint, we are making certain independence assumptions about our random variables. This makes more precise a little, anyway the difference between the family specified by the graphical model and the family of all joints. Q: Are these the only independence assumptions we are making? The Bayes ball algorithm and dseparation Note that a node s parents separate it from its ancestors. It appears that conditional independencies are related to the graph and, in particular, to graph separation. We will next uncover the relationship between graph separation and conditional independence. To do this, and deepen our understanding of independence and graphical models, we look at three simple graphs. We ask which independencies hold in these graphs and consider the relationship to classical graph separation. The first is a little sequence, X Y Z p.x; y; z/ D p.x/p.y j x/p.z j y/: (20) Here, X? Z j Y: (21) To see this, p.x; y; z/ p.x j y; z/ D p.y; z/ (22) p.x/p.y j x/p.z j y/ D p.z j y/ P x p.x 0 /p.y j x 0 / 0 (23) p.x; y/ D p.y/ (24) D p.x j y/ (25) We assert that no other independencies hold. (E.g., is it not true that X? Z.) 5
6 Here we do not need to do this algebra. This is one of the basic conditional independencies. The parent of Z is Y, and X is a nonparent ancestor of Z. Important subtlety: This means that other independencies do not necessarily hold. For some settings of p.y j x/ it may be true that X? Z. But, not for all. In other words, a more restrictive family of joints will be contained in the less restrictive family. In this graph, conditional independence can be interpreted as graph separation. In graphical models notation, we shade the node that we are conditioning on: We can see that Y separates X and Z. X Y Z. The intuition: X is the past, Y is the present, Z is the future. Given the present, the past is independent of the future. This is the Markov assumption. This graph is a three step Markov chain. The second graph is a little tree, Y X Z p.x; y; z/ D p.y/p.x j y/p.z j y/ (26) Here we have again that X? Z j Y. We calculate that the conditional joint factorizes, p.y/p.x j y/p.z j y/ p.x; z j y/ D p.y/ (27) D p.x j y/p.z j y/ (28) We assert that no other conditional independencies hold. Again, simple graph separation indicates independence, 6
7 Y X Z The intuition behind this graph comes from a latent variable model. In our previous lecture, this graph describes the unknown coin flipped twice. As another example, let X be shoe size and Z be amount of gray hair. In general, these are dependent variables. But suppose Y is age. Conditioned on Y, X and Z become independent. Graphically, we can see this. It is through age that shoe size and gray hair depend on each other. The last simple graph is an inverse tree X Z Y p.x; y; z/ D p.x/p.z/p.y j x; z/ (29) Here the only independence statement is X? Z. In particular, it is not necessarily true that X? Z j Y. For intuition, think of a causal model: Y is I m late for lunch ; X is I m abducted by aliens, a possible cause of being late; Z is My watch is broken, another possible cause. Marginally, being abducted and breaking my watch are independent. But conditioned on my lateness, knowing about one tells us about the likelihood of the other. (E.g., if I m late and you know that my watch is broken, then this decreases the chance that I was abducted.) Alas this independency does not correspond to graph separation. With these simple graphs in hand, we can now discuss dseparation, a notion of graph separability that lets us determine the validity of any conditional independence statement in a directed graphical model. Suppose we are testing a conditional independence statement, X A? X B j X C : (30) 7
8 We shade the nodes being conditioned on. We then decide, using the Bayes ball algorithm, whether the conditioned nodes dseparate the nodes on either side of the independence relation. The Bayes ball algorithm is a reachability algorithm. We start balls off at one of the sets of variables. If they can reach one of the other set then the conditional independence statement is false. The balls bounce around the graph according to rules based on the three simple graphs. We consider a ball starting at X and going through Y on its way to Z. (To be clear, if the move is allowed, then the next step is for the ball to be at Y and we ask if it can go through Z en route to another node.) Note that it does not matter if the source node X and destination node Z are shaded. Here are the rules: In addition, there are rules derived by contemplating a ball going through a node and then back to the source node: Some examples: 1. Look at our example graph. 8
9 (a) X 1? X 6 j fx 2 ; X 3 g? Yes. (b) X 2? X 3 j fx 1 ; X 6 g? No. 2. A Markov chain is the simple sequence graph with any length sequence. The basic conditional independencies are that But Bayes ball tells us more, e.g. X ic1? X 1W.i 1/ j X i : (31) X 1? X 5 j X 4 X 1? X 5 j X 2 X 1? X 5 j X 2 ; X 4 3. Now consider a hidden Markov model which is used, for example, in speech recognition. The Bayes ball algorithm reveals that there are no conditional independencies among the observations. 4. (Optional) Look at a tree model, e.g., of genetic sequences in a family tree. What kinds of independencies do we see? 5. Look at a Bayesian hierarchical regression model. (E.g., consider testing in different schools.) How are the groups related? What if we know the prior? Remarks on Bayes ball: It s not an algorithm that is necessarily very interesting to implement. But it s very useful to look at graphs i.e., at structured joint distributions and understand the complete set of conditional independence and independence assumptions that are being made. As we have shown, this is not obvious either from the joint distribution or the structure alone. The idea of a ball bouncing around is a theme that we will come back to. It won t be balls, but be messages (i.e., information). Just as balls bouncing around the graph help us understand independence, messages traveling on the graph will help us make probabilistic computations. The HammersleyClifford theorem Punchline: Consider two families of joint probability distributions, both obtained from the graphical model G. 1. Family of joints found by ranging over all conditional probability tables associated with G. 2. All joints that respect all conditional independence statements, implied by G and dseparation. 9
10 The HammersleyClifford theorem says that these families are the same. More verbose: We find the first family by varying the local conditional probability tables and computing the resulting joint from its factorization. This is what we meant earlier when we said that a graphical model defines a family of probability distributions. We obtain the second family as follows. First, compute every conditional independence statement that is implied by the graph. (Use Bayes ball.) Then, consider every joint distribution of the same set of variables. Note this does not reference the local conditional probability tables. For each joint, check whether all the conditional independence statements hold. If one does not, throw the joint away. Those that remain are the second family. The HammersleyClifford theorem says that these two families of joints one obtained by checking conditional independencies and the other obtained by varying local probability tables are the same. As stated in the chapter, this theorem is at the core of the graphical models formalism. It makes precise and clear what limitations (or assumptions) we place on the family of joint distributions when we specify a graphical model. Undirected graphical models In this class, we will mainly focus on directed graphical models. However, undirected graphical models, which are also known as Markov random fields, are a very useful formalism as well. They are important to learn about to be fluent in graphical models, will be useful later when we talk about exact inference, and further refine the picture of the relationship between graphs and probability models. A definition, via conditional independencies When discussing directed models, we began with a definition of how to map a graph to a joint and then showed how the resulting conditional independencies can be seen from the graph. Here we will go in reverse. Based on an undirected graph, we will first define the conditional independencies that we want to hold in its corresponding joint distribution. We will then define the form of that distribution. Consider an undirected graph G D.V; E/ and three sets of nodes A, B, and C. We will want a joint distribution such that X A? X C j X B if X B separates X A and X C, in the usual graphtheoretic sense of separate. Formally, quoting from the book, if every path from a node in X_A to a node in X_C includes at least one node in X_B then we assert that X A? X C j X B. 10
11 Again we emphasize that we are representing a family of distributions. These are the conditional independence statements that (we assert) have to hold. For various instantiations of the graphical model (i.e., various members of the family) other conditional independencies may also hold. Undirected and directed graphical models are different Consider the families of families expressable by directed and undirected graphical models, respectively. Not all directed graphical models can be written as an undirected graphical model, and vice versa. First consider this directed graph, X Z Y As we said, the only conditional independence statement that is true for this graph is X? Z. We cannot write an undirected graphical model such that this is the only conditional independence statement, i.e. where X 6? Z j Y. Now consider this undirected graph, [Four nodes connected in a square] This graph expresses the family characterized by the following independencies, And these are the only ones. X? Y j fw; Zg (32) W? Z j fx; Y g (33) We cannot write down a directed graphical model such that these are the only two conditional independence statements. Exercise: Confirm this. Note that there are types of directed and undirected graphical models that can be written as either. We will one such important class when we talk about inference. But, in general, we have just demonstrated that they have different expressive power. The joint distribution in an undirected graphical model From the conditional independencies, we will now develop a representation of the joint distribution. Our goal is to represent the joint as a product of local functions, which we 11
12 will call potential functions, whose arguments are subsets of the random variables, p.x 1 ; : : : ; x n / D 1 Z Y S2S.x S / (34) Here S is a set of nodes,./ are arbitrary different potential functions (notation overloaded), and S is a collection of subsets. (We specify them later.) We would like these functions to be nonnegative but otherwise arbitrary, so we will be satisfied with specifying them up to a scaling constant Z. (We use this notation to be consistent with the literature; Z is not a random variable.) This is called the normalizing constant, and will be an important quantity for much of this course. We need to define what we mean by local. This amounts to choosing the arguments of each of the potential functions, i.e., choosing the subsets S. Let s return to the conditional independencies that we are hoping to assume with this representation. These imply that if two nodes X 1 and X 3 are separated by a third X 2 then X 1? X 3 j X 2. This implies that the conditional distribution factorizes, p.x 1 ; x 3 j x 2 / D p.x 1 j x 2 /p.x 3 j x 2 /: (35) This further implies that the three nodes cannot participate in a single potential. Why? If there were an arbitrary potential function.x 1 ; x 2 ; x 3 / in the joint distribution of Equation (34) then it would be impossible for the conditional (which, recall, is proportional to the joint) to factorize across x 1 and x 3. Maximal cliques. This is suggestive that the potential functions should only be defined on cliques, which are sets of nodes that are fully connected, or subsets of cliques. Because of the direct connections between all nodes in a clique we are guaranteed that the graph does not imply any conditional independencies between them; thus it is safe to include them in the arguments to a potential. Conversely, as we argued above, if two nodes are not directly connected in the graph then there is a conditional independence statement that we can make. Thus, they should not appear together in a potential. In the theory around undirected graphical models, the joint is defined on the set of maximal cliques, i.e., completely connected components of the graph that cannot be expanded without breaking complete connectedness. Every node is part of a maximal clique. Thus, we can write the joint as p.x/ D 1 Z Y C 2C.x C /: (36) Here, C is the set of maximal cliques and C is a particular clique (i.e., set of nodes). The normalizing constant is Z D X Y.x C /: (37) x It is difficult to compute. (Why?) We ll come back to that later in the semester. C 2C 12
13 This joint distribution respects the set of conditional independence statements implied by usual graph separability on the underlying graph. Finally, in practice we often define undirected graphical models in terms of other cliques, in addition to or instead of maximal cliques. As long as we don t steer beyond a maximal clique, this preserves the relationship between graph separation and conditional independence. Interpreting potentials. The potential functions we set up are arbitrary positive valued functions. They are not conditional probabilities (necessarily) as in the directed graphical models case. However, they can be interpreted as providing agreement to configurations of variables that have high probability. If the potential on.x 1 ; x 2 / is high then the configuration with those values has higher probability. Though we will not discuss it in depth, this is how undirected graphical models play a large role in statistical physics (the field in which they were invented). HammersleyClifford for undirected graphical models. We can state a similar theorem for undirected graphical models as we did for directed graphical models. Fix an undirected graph G. Define one family of distributions by ranging over all possible potential functions over the maximal cliques of the graph, and calculating the joint distribution in Equation (36). Define a second family of distributions by looking at all joint distributions over the set of nodes in the graph and filtering out only those for which the set of conditional independence statements defined by graph separability holds. These two sets of distributions are the same. 13
Gibbs Sampling and Online Learning Introduction
Statistical Techniques in Robotics (16831, F14) Lecture#10(Tuesday, September 30) Gibbs Sampling and Online Learning Introduction Lecturer: Drew Bagnell Scribes: {Shichao Yang} 1 1 Sampling Samples are
More informationCourse: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.
More informationDiscrete Structures for Computer Science
Discrete Structures for Computer Science Adam J. Lee adamlee@cs.pitt.edu 6111 Sennott Square Lecture #20: Bayes Theorem November 5, 2013 How can we incorporate prior knowledge? Sometimes we want to know
More information5 Directed acyclic graphs
5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate
More information3. The Junction Tree Algorithms
A Short Course on Graphical Models 3. The Junction Tree Algorithms Mark Paskin mark@paskin.org 1 Review: conditional independence Two random variables X and Y are independent (written X Y ) iff p X ( )
More informationLecture 16 : Relations and Functions DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/29/2011 Lecture 16 : Relations and Functions Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT In Lecture 3, we described a correspondence
More informationCHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
More informationLecture 7: NPComplete Problems
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NPComplete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit
More informationYou know from calculus that functions play a fundamental role in mathematics.
CHPTER 12 Functions You know from calculus that functions play a fundamental role in mathematics. You likely view a function as a kind of formula that describes a relationship between two (or more) quantities.
More information6.3 Conditional Probability and Independence
222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted
More informationQuestion 2 Naïve Bayes (16 points)
Question 2 Naïve Bayes (16 points) About 2/3 of your email is spam so you downloaded an open source spam filter based on word occurrences that uses the Naive Bayes classifier. Assume you collected the
More informationCORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREERREADY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREERREADY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
More informationLecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs
CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like
More informationCell Phone based Activity Detection using Markov Logic Network
Cell Phone based Activity Detection using Markov Logic Network Somdeb Sarkhel sxs104721@utdallas.edu 1 Introduction Mobile devices are becoming increasingly sophisticated and the latest generation of smart
More informationCost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:
CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm
More informationChapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a realvalued function defined on the sample space of some experiment. For instance,
More informationOutline. NPcompleteness. When is a problem easy? When is a problem hard? Today. Euler Circuits
Outline NPcompleteness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2pairs sum vs. general Subset Sum Reducing one problem to another Clique
More informationBayesian networks  Timeseries models  Apache Spark & Scala
Bayesian networks  Timeseries models  Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup  November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
More informationGraphical Models, Exponential Families, and Variational Inference
Foundations and Trends R in Machine Learning Vol. 1, Nos. 1 2 (2008) 1 305 c 2008 M. J. Wainwright and M. I. Jordan DOI: 10.1561/2200000001 Graphical Models, Exponential Families, and Variational Inference
More informationDecision Trees and Networks
Lecture 21: Uncertainty 6 Today s Lecture Victor R. Lesser CMPSCI 683 Fall 2010 Decision Trees and Networks Decision Trees A decision tree is an explicit representation of all the possible scenarios from
More informationTagging with Hidden Markov Models
Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Partofspeech (POS) tagging is perhaps the earliest, and most famous,
More informationSupervised Learning (Big Data Analytics)
Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used
More informationMessagepassing sequential detection of multiple change points in networks
Messagepassing sequential detection of multiple change points in networks Long Nguyen, Arash Amini Ram Rajagopal University of Michigan Stanford University ISIT, Boston, July 2012 Nguyen/Amini/Rajagopal
More informationApplied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
More informationStudy Manual. Probabilistic Reasoning. Silja Renooij. August 2015
Study Manual Probabilistic Reasoning 2015 2016 Silja Renooij August 2015 General information This study manual was designed to help guide your self studies. As such, it does not include material that is
More information8.1 Min Degree Spanning Tree
CS880: Approximations Algorithms Scribe: Siddharth Barman Lecturer: Shuchi Chawla Topic: Min Degree Spanning Tree Date: 02/15/07 In this lecture we give a local search based algorithm for the Min Degree
More informationLecture 10: Regression Trees
Lecture 10: Regression Trees 36350: Data Mining October 11, 2006 Reading: Textbook, sections 5.2 and 10.5. The next three lectures are going to be about a particular kind of nonlinear predictive model,
More information[Refer Slide Time: 05:10]
Principles of Programming Languages Prof: S. Arun Kumar Department of Computer Science and Engineering Indian Institute of Technology Delhi Lecture no 7 Lecture Title: Syntactic Classes Welcome to lecture
More informationInformation Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture  17 ShannonFanoElias Coding and Introduction to Arithmetic Coding
More information1. Nondeterministically guess a solution (called a certificate) 2. Check whether the solution solves the problem (called verification)
Some N P problems Computer scientists have studied many N P problems, that is, problems that can be solved nondeterministically in polynomial time. Traditionally complexity question are studied as languages:
More information2. (a) Explain the strassen s matrix multiplication. (b) Write deletion algorithm, of Binary search tree. [8+8]
Code No: R05220502 Set No. 1 1. (a) Describe the performance analysis in detail. (b) Show that f 1 (n)+f 2 (n) = 0(max(g 1 (n), g 2 (n)) where f 1 (n) = 0(g 1 (n)) and f 2 (n) = 0(g 2 (n)). [8+8] 2. (a)
More informationThe sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].
Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real
More informationAnalysis of Algorithms, I
Analysis of Algorithms, I CSOR W4231.002 Eleni Drinea Computer Science Department Columbia University Thursday, February 26, 2015 Outline 1 Recap 2 Representing graphs 3 Breadthfirst search (BFS) 4 Applications
More informationLoad Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: liuxh@systems.stanford.edu Abstract Load
More informationBasic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011
Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011 A. Miller 1. Introduction. The definitions of metric space and topological space were developed in the early 1900 s, largely
More informationThe Chinese Restaurant Process
COS 597C: Bayesian nonparametrics Lecturer: David Blei Lecture # 1 Scribes: Peter Frazier, Indraneel Mukherjee September 21, 2007 In this first lecture, we begin by introducing the Chinese Restaurant Process.
More informationGerry Hobbs, Department of Statistics, West Virginia University
Decision Trees as a Predictive Modeling Method Gerry Hobbs, Department of Statistics, West Virginia University Abstract Predictive modeling has become an important area of interest in tasks such as credit
More informationA NonLinear Schema Theorem for Genetic Algorithms
A NonLinear Schema Theorem for Genetic Algorithms William A Greene Computer Science Department University of New Orleans New Orleans, LA 70148 bill@csunoedu 5042806755 Abstract We generalize Holland
More informationLargeSample Learning of Bayesian Networks is NPHard
Journal of Machine Learning Research 5 (2004) 1287 1330 Submitted 3/04; Published 10/04 LargeSample Learning of Bayesian Networks is NPHard David Maxwell Chickering David Heckerman Christopher Meek Microsoft
More informationEx. 2.1 (Davide Basilio Bartolini)
ECE 54: Elements of Information Theory, Fall 00 Homework Solutions Ex.. (Davide Basilio Bartolini) Text Coin Flips. A fair coin is flipped until the first head occurs. Let X denote the number of flips
More informationBayesian Tutorial (Sheet Updated 20 March)
Bayesian Tutorial (Sheet Updated 20 March) Practice Questions (for discussing in Class) Week starting 21 March 2016 1. What is the probability that the total of two dice will be greater than 8, given that
More informationSocial Media Mining. Graph Essentials
Graph Essentials Graph Basics Measures Graph and Essentials Metrics 2 2 Nodes and Edges A network is a graph nodes, actors, or vertices (plural of vertex) Connections, edges or ties Edge Node Measures
More informationWHERE DOES THE 10% CONDITION COME FROM?
1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationLecture 8 The Subjective Theory of Betting on Theories
Lecture 8 The Subjective Theory of Betting on Theories Patrick Maher Philosophy 517 Spring 2007 Introduction The subjective theory of probability holds that the laws of probability are laws that rational
More informationImportant Probability Distributions OPRE 6301
Important Probability Distributions OPRE 6301 Important Distributions... Certain probability distributions occur with such regularity in reallife applications that they have been given their own names.
More informationGraph Theory Problems and Solutions
raph Theory Problems and Solutions Tom Davis tomrdavis@earthlink.net http://www.geometer.org/mathcircles November, 005 Problems. Prove that the sum of the degrees of the vertices of any finite graph is
More informationA Second Course in Mathematics Concepts for Elementary Teachers: Theory, Problems, and Solutions
A Second Course in Mathematics Concepts for Elementary Teachers: Theory, Problems, and Solutions Marcel B. Finan Arkansas Tech University c All Rights Reserved First Draft February 8, 2006 1 Contents 25
More informationBayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
More informationU.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009. Notes on Algebra
U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009 Notes on Algebra These notes contain as little theory as possible, and most results are stated without proof. Any introductory
More informationConditional Probability, Independence and Bayes Theorem Class 3, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom
Conditional Probability, Independence and Bayes Theorem Class 3, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Know the definitions of conditional probability and independence
More informationMaster s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
More informationChapter ML:IV. IV. Statistical Learning. Probability Basics Bayes Classification Maximum aposteriori Hypotheses
Chapter ML:IV IV. Statistical Learning Probability Basics Bayes Classification Maximum aposteriori Hypotheses ML:IV1 Statistical Learning STEIN 20052015 Area Overview Mathematics Statistics...... Stochastics
More informationA Sublinear Bipartiteness Tester for Bounded Degree Graphs
A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublineartime algorithm for testing whether a bounded degree graph is bipartite
More informationIntroduction to computer science
Introduction to computer science Michael A. Nielsen University of Queensland Goals: 1. Introduce the notion of the computational complexity of a problem, and define the major computational complexity classes.
More informationData Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1
Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2011 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields
More informationPenalized regression: Introduction
Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20thcentury statistics dealt with maximum likelihood
More informationLecture 1: Course overview, circuits, and formulas
Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik
More informationComputational complexity theory
Computational complexity theory Goal: A general theory of the resources needed to solve computational problems What types of resources? Time What types of computational problems? decision problem Decision
More informationLecture Note 1 Set and Probability Theory. MIT 14.30 Spring 2006 Herman Bennett
Lecture Note 1 Set and Probability Theory MIT 14.30 Spring 2006 Herman Bennett 1 Set Theory 1.1 Definitions and Theorems 1. Experiment: any action or process whose outcome is subject to uncertainty. 2.
More informationSocial Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
More informationMarkov random fields and Gibbs measures
Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).
More information2.2. Instantaneous Velocity
2.2. Instantaneous Velocity toc Assuming that your are not familiar with the technical aspects of this section, when you think about it, your knowledge of velocity is limited. In terms of your own mathematical
More informationChapter 7: Products and quotients
Chapter 7: Products and quotients Matthew Macauley Department of Mathematical Sciences Clemson University http://www.math.clemson.edu/~macaule/ Math 42, Spring 24 M. Macauley (Clemson) Chapter 7: Products
More informationAnswer Key for California State Standards: Algebra I
Algebra I: Symbolic reasoning and calculations with symbols are central in algebra. Through the study of algebra, a student develops an understanding of the symbolic language of mathematics and the sciences.
More informationBayesian Networks. Read R&N Ch. 14.114.2. Next lecture: Read R&N 18.118.4
Bayesian Networks Read R&N Ch. 14.114.2 Next lecture: Read R&N 18.118.4 You will be expected to know Basic concepts and vocabulary of Bayesian networks. Nodes represent random variables. Directed arcs
More informationDistributed Computing over Communication Networks: Maximal Independent Set
Distributed Computing over Communication Networks: Maximal Independent Set What is a MIS? MIS An independent set (IS) of an undirected graph is a subset U of nodes such that no two nodes in U are adjacent.
More informationProblem of the Month Through the Grapevine
The Problems of the Month (POM) are used in a variety of ways to promote problem solving and to foster the first standard of mathematical practice from the Common Core State Standards: Make sense of problems
More informationHidden Markov Models
8.47 Introduction to omputational Molecular Biology Lecture 7: November 4, 2004 Scribe: HanPang hiu Lecturer: Ross Lippert Editor: Russ ox Hidden Markov Models The G island phenomenon The nucleotide frequencies
More informationBargaining Solutions in a Social Network
Bargaining Solutions in a Social Network Tanmoy Chakraborty and Michael Kearns Department of Computer and Information Science University of Pennsylvania Abstract. We study the concept of bargaining solutions,
More information1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let
Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as
More informationThe UnionFind Problem Kruskal s algorithm for finding an MST presented us with a problem in datastructure design. As we looked at each edge,
The UnionFind Problem Kruskal s algorithm for finding an MST presented us with a problem in datastructure design. As we looked at each edge, cheapest first, we had to determine whether its two endpoints
More informationIf A is divided by B the result is 2/3. If B is divided by C the result is 4/7. What is the result if A is divided by C?
Problem 3 If A is divided by B the result is 2/3. If B is divided by C the result is 4/7. What is the result if A is divided by C? Suggested Questions to ask students about Problem 3 The key to this question
More informationFairfield Public Schools
Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity
More informationChapter 3. Distribution Problems. 3.1 The idea of a distribution. 3.1.1 The twentyfold way
Chapter 3 Distribution Problems 3.1 The idea of a distribution Many of the problems we solved in Chapter 1 may be thought of as problems of distributing objects (such as pieces of fruit or pingpong balls)
More informationMA107 Precalculus Algebra Exam 2 Review Solutions
MA107 Precalculus Algebra Exam 2 Review Solutions February 24, 2008 1. The following demand equation models the number of units sold, x, of a product as a function of price, p. x = 4p + 200 a. Please write
More information1 Local Brouwer degree
1 Local Brouwer degree Let D R n be an open set and f : S R n be continuous, D S and c R n. Suppose that the set f 1 (c) D is compact. (1) Then the local Brouwer degree of f at c in the set D is defined.
More informationThe Two Envelopes Problem
1 The Two Envelopes Problem Rich Turner and Tom Quilter The Two Envelopes Problem, like its better known cousin, the Monty Hall problem, is seemingly paradoxical if you are not careful with your analysis.
More informationHow To Check For Differences In The One Way Anova
MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. OneWay
More informationBasics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
More informationData Mining on Social Networks. Dionysios Sotiropoulos Ph.D.
Data Mining on Social Networks Dionysios Sotiropoulos Ph.D. 1 Contents What are Social Media? Mathematical Representation of Social Networks Fundamental Data Mining Concepts Data Mining Tasks on Digital
More informationExam Introduction Mathematical Finance and Insurance
Exam Introduction Mathematical Finance and Insurance Date: January 8, 2013. Duration: 3 hours. This is a closedbook exam. The exam does not use scrap cards. Simple calculators are allowed. The questions
More informationIntroduction to Markov Chain Monte Carlo
Introduction to Markov Chain Monte Carlo Monte Carlo: sample from a distribution to estimate the distribution to compute max, mean Markov Chain Monte Carlo: sampling using local information Generic problem
More informationarxiv:1112.0829v1 [math.pr] 5 Dec 2011
How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly
More informationStatistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics: Behavioural
More informationTopologybased network security
Topologybased network security Tiit Pikma Supervised by Vitaly Skachek Research Seminar in Cryptography University of Tartu, Spring 2013 1 Introduction In both wired and wireless networks, there is the
More informationFactor Graphs and the SumProduct Algorithm
498 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 47, NO. 2, FEBRUARY 2001 Factor Graphs and the SumProduct Algorithm Frank R. Kschischang, Senior Member, IEEE, Brendan J. Frey, Member, IEEE, and HansAndrea
More informationTHE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear
More informationAttribution. Modified from Stuart Russell s slides (Berkeley) Parts of the slides are inspired by Dan Klein s lecture material for CS 188 (Berkeley)
Machine Learning 1 Attribution Modified from Stuart Russell s slides (Berkeley) Parts of the slides are inspired by Dan Klein s lecture material for CS 188 (Berkeley) 2 Outline Inductive learning Decision
More informationPart III: Machine Learning. CS 188: Artificial Intelligence. Machine Learning This Set of Slides. Parameter Estimation. Estimation: Smoothing
CS 188: Artificial Intelligence Lecture 20: Dynamic Bayes Nets, Naïve Bayes Pieter Abbeel UC Berkeley Slides adapted from Dan Klein. Part III: Machine Learning Up until now: how to reason in a model and
More informationBNG 202 Biomechanics Lab. Descriptive statistics and probability distributions I
BNG 202 Biomechanics Lab Descriptive statistics and probability distributions I Overview The overall goal of this short course in statistics is to provide an introduction to descriptive and inferential
More information1 Maximum likelihood estimation
COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N
More informationThe Goldberg Rao Algorithm for the Maximum Flow Problem
The Goldberg Rao Algorithm for the Maximum Flow Problem COS 528 class notes October 18, 2006 Scribe: Dávid Papp Main idea: use of the blocking flow paradigm to achieve essentially O(min{m 2/3, n 1/2 }
More informationChapter 28. Bayesian Networks
Chapter 28. Bayesian Networks The Quest for Artificial Intelligence, Nilsson, N. J., 2009. Lecture Notes on Artificial Intelligence, Spring 2012 Summarized by Kim, ByoungHee and Lim, ByoungKwon Biointelligence
More informationLAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE
LAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE MAT 119 STATISTICS AND ELEMENTARY ALGEBRA 5 Lecture Hours, 2 Lab Hours, 3 Credits Pre
More informationConductance, the Normalized Laplacian, and Cheeger s Inequality
Spectral Graph Theory Lecture 6 Conductance, the Normalized Laplacian, and Cheeger s Inequality Daniel A. Spielman September 21, 2015 Disclaimer These notes are not necessarily an accurate representation
More informationDATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More informationProtein Protein Interaction Networks
Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks YoungRae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics
More informationparent ROADMAP MATHEMATICS SUPPORTING YOUR CHILD IN HIGH SCHOOL
parent ROADMAP MATHEMATICS SUPPORTING YOUR CHILD IN HIGH SCHOOL HS America s schools are working to provide higher quality instruction than ever before. The way we taught students in the past simply does
More informationLearning from Data: Naive Bayes
Semester 1 http://www.anc.ed.ac.uk/ amos/lfd/ Naive Bayes Typical example: Bayesian Spam Filter. Naive means naive. Bayesian methods can be much more sophisticated. Basic assumption: conditional independence.
More information