counting principles, including permutations and ......the binomial theorem: ... presentation of...
TRANSCRIPT
1
Counting principles, including permutations and combinations.
The binomial theorem: expansion of π + π π, π πΊ π΅.
β THE PRODUCT RULE
If there are π different ways of performing an operation and for each of these there are π different ways of performing a second independent operation, then there are ππ different ways of performing the two operations in succession. The product principle can be extended to three or more successive operations. The number of different ways of performing an operation is equal to the sum of the different mutually exclusive possibilities.
β COUNTING PATHS
The word πππ suggests multiplying the possibilities The word ππ suggests adding the possibilities. If the order doesn't matter, it is a Combination. If the order does matter it is a Permutation.
β PERMUTATIONS (order matters)
A permutation of a group of symbols is any arrangement of those symbols in a definite order.
β Permutations of π different object : π!
Explanation: Assume you have n different symbols and therefore n places to fill in your arrangement. For the first place, there are n different possibilities. For the second place, there are n β 1 possible symbols, β¦ until we saturate all the places. According to the product principle, therefore, we have n (n β 1)(n β 2)(n β 3)β―1 different arrangements, or n!
Wise Advice: If a group of items have to be kept together, treat the items as one object. Remember that there may
be permutations of the items within this group too. β Permutations of π different objects out of π different available (no repetition allowed) :
πππ =
π!
π β π != π β π β 1 βββ π β π + 1
Good logic to apply to similar questions straightforward: Suppose we have 10 letters and want to make groups of 4 letters. For four-letter permutations, there are 10 possibilities for the first letter, 9 for the second, 8 for the third, and 7 for the last letter. We can find the total number of different four-letter permutations by multiplying 10 x 9 x 8 x 7 = 5040.
β Permutations with repetition of π different objects out of π different available = ππ
There are n possibilities for the first choice, THEN there are n possibilities for the second choice, and so on, multiplying each time.)
β COMBINATIONS (order doesnβt matters)
It is the number of ways of choosing π objects out of π available given that βͺ The order of the elements does not matter. βͺ The elements are not repeated [such as lottery numbers (2,14,15,27,30,33)]
The easiest way to explain it is to:
assume that the order does matter (ie permutations),
then alter it so the order does not matter. Since the combination does not take into account the order, we have to divide the permutation of the total number of symbols available by the number of redundant possibilities. π selected objects have a number of redundancies equal to the permutation of the objects π! (since order doesnβt matter)
2
However, we also need to divide the permutation n! by the permutation of the objects that are not selected, that is to say π β π ! .
π!
π! π β π !
(ππ) β‘ πͺπ
π β‘ πͺππ =
π!
π! π β π !=
π π β π π β π π β π + π
π!
β Binomial Expansion/Theorem
π + π π = β (ππ) ππβπππ = ππ + (
π1) ππβ1π + β―+ (
ππ) ππβπππ + β―+ ππ
π
π=0
β Binomial Coefficient
(ππ) is the coefficient of the term containing ππβπππ in the expansion of π + π π
(ππ) =
π π β 1 π β 2 β― π β π + 1
π!=
π!
π! π β π !=
π!
π β π ! π!= (
ππ β π
)
πβπ πππππππ π‘πππ ππ π + 1 π‘β π‘πππ ππ : ππ+1 = (ππ)ππβπππ
The constant term is the term containing no variables.
When finding the coefficient of π₯π always consider the set of all terms containing π₯π
Probability The number of trials is the total number of times the βexperimentβ is repeated.
The outcomes are the different results possible for one trial of the experiment.
Equally likely outcomes are expected to have equal frequencies.
The sample space, U, is the set of all possible outcomes of an experiment.
And event is the occurrence of one particular outcome.
π π΄ =π π΄
π π
Complementary Events
Two events are described as complementary if they are the only two possible outcomes.
Two complementary events are mutually exclusive.
Since an event must either occur or not occur, the probability of the event either occurring or not
occurring must be 1.
π· π¨ + π· π¨β² = π
Use when you need probability that an event will not happen
Possibility when we are interested in more than one outcome (events are βandβ, βorβ, βat leastβ)
P(A) is the probability of an event A occurring in one trial,
n(A) is the number of times event A occurs in the sample space n(U) is
the total number of possible outcomes.
π π π (π0) β‘ 1 0! β‘ 1
3
Combined Events
βͺ π’ππππ β‘ πππ‘βππ β© πππ‘πππ πππ‘πππ β‘ πππ‘β/πππ
Given two events, B and A, the probability of at least one of the two events occurring,
π π΄ βͺ π΅ = π π΄ + π π΅ β π π΄ β© π΅
πΌπ‘ ππ ππππππ‘πππ‘ π‘π ππππ€ βππ€ π‘π πππ‘ π π΄ β© π΅
For mutually exclusive events (no possibility that A and B occurring at the same time)
Turning left and turning right (you can't do both at the same time)
Tossing a coin: Heads and Tails
π π΄ βͺ π΅ = π π΄ + π π΅ π π΄ β© π΅ = β
For non - mutually exclusive we are going to find conditional probability for
Independent and Dependent Events
A bag contains three different kinds of marbles: red, blue and green. You pick the marble twice. Probability of picking up the red one (or any) the second time depends weather you put back the first marble or not.
β’ Independent Events: β’ Dependent Events:
the probability that one event occurs probability of one event occurring influences in no way affects the probability of the likelihood of the other event the other event occurring. You put the first marble back You donβt put the first marble
β Conditional Probability:
Given two events, B and A, the conditional probability of an event A is the probability that the event will occur given the knowledge that an event B has already occurred. This probability is written as (notation for the probability of A given B) P (A|B )
Probability of the intersection of A and B (both events occur) is: π π΄ β© π΅ = π π΅ π π΄|π΅
β’ Independent Events: β’ Dependent Events:
π π΄|π΅ = π π΄ = π π΄|π΅β² π π΄ β© π΅ = π π΅ π π΄|π΅
π΄ ππππ πππ‘ ππππππ ππ π΅ πππ ππ π΅β² π π΄|π΅ πππππ’πππ‘ππ πππππππππ ππ π‘βπ ππ£πππ‘ π΅
π π΄ β© π΅ = π π΄ π π΅ π π΄ β© π΅ = π π΅ π π΄|π΅
π π΄|π΅ =π π΄ β© π΅
π π΅
either A or B or both
P(A) includes part of B from intersection P(B) includes part of A from intersection
π π΄ β© π΅ (both A and B) was counted twice, so one has to be subtracted
4
β Use of Venn diagrams, tree diagrams and tables of outcomes to solve problems.
1. Venn Diagrams
The probability is found using the principle π π΄ =π π΄
π π
2. Tree diagrams
A more flexible method for finding probabilities is known as a tree diagram.
β§ͺ Bayesβ Theorem
π π΄ β© π΅ = π π΅ π π΄|π΅ βΉ π π΄ β© π΅ = π π΄ π π΅|π΄
π π΄|π΅ = π π΄ β© π΅
π π΅ =
π π΄ π π΅|π΄
π π΅ π΅ππ¦ππ β² π‘βπππππ
βͺ Another form of Bayesβ theorem (Formula booklet)
From tree diagram:
there are two ways to get A, either after B has happen or after B has not happened:
π π΄ = π π΅ π π΄|π΅ + π π΅β² π π΄|π΅β² βΉ π π΅|π΄ =π π΅ π π΄|π΅
π π΅ π π΄|π΅ + π π΅β² π π΄|π΅β²
This allows one to calculate the probabilities
of the occurrence of events, even where trials
are non-identical (where π π΄|π΄ β π π΄ ),
through the product principle.
5
βͺ Extension of Bayesβ Theorem
If there are more options than simply B occurs or B doesnβt occur, for example if there were three possible outcomes for the first event B1, B2, and B3
Probability of A occurring is: π π΅1 π π΄|π΅1 + π π΅2 π π΄|π΅2 + π π΅3 π π΄|π΅3
π π΅π|π΄ =π π΅π π π΄|π΅π
π π΅1 π π΄|π΅1 + π π΅2 π π΄|π΅2 + π π΅3 π π΄|π΅3
Outcomes B1, B2, and B3 must cover all the possible outcomes.
Descriptive Statistics
Concepts of population, sample, random sample and frequency distribution of discrete and continuous data.
A population is the set of all individuals with a given value for a variable associated with them.
A sample is a small group of individuals randomly selected (in the case of a random sample) from the population as a whole, used as a representation of the population as a whole.
The frequency distribution of data is the number of individuals within a sample or population for each value of the associated variable in discrete data, or for each range of values for the associated variable in continuous data.
new guidelines in IB MATH: population β‘ sample
Presentation of data: frequency tables and diagrams Grouped data: mid-interval values, interval width, upper and lower interval boundaries, frequency histograms.
Mid interval values are found by halving the difference between the upper and lower interval boundaries.
The interval width is simply the distance between the upper and lower interval boundaries.
Frequency histograms are drawn with interval width proportional to bar width and frequency as the height.
Median, mode; quartiles, percentiles.
Range; interquartile range; variance, standard deviation.
Mode (discrete data) is the most frequently occurring value in the data set.
Modal class (continuous data) is the most frequently occurring class.
Median is the middle value of an ordered data set.
For an odd number of data, the median is middle data.
For an even number of data, the median is average of two middle data.
6
Percentile is the score bellow which a certain percentage of the data lies.
Lower quartile (Q1) is the 25th percentile.
Median (Q2) is the 50th percentile.
Upper quartile (Q3) is the 75th percentile.
Range is the difference between the highest and lowest value in the data set.
The interquartile range is Q3βQ1.
Cumulative frequency is the frequency of all values less than a given value.
The population mean, ΞΌ is generally unknown but the sample mean, οΏ½Μ οΏ½ used to serve as an unbiased estimate of this mean. That used to be. From now on for the examination purposes, data will be treated as the population. Estimation of mean and variance of population from a sample is no longer required.
Discrete and Continuous Random Variables
A variable X whose value depends on the outcome of a random process is called a random variable. For any random variable there is a probability distribution/ function associated with it.
β Probability distribution/ function
Discrete Random Variables
P(X = x), the probability distribution of x, involves listing P(π₯π ) for each π₯π .
1. 0 β€ π π = π₯ β€ 1
2. βπ π = π₯ = 1
π₯
3. π π = π₯π = 1 β β π π = π₯π
πβ π
[π ππ£πππ‘ π₯π ππππ’ππ = 1 β π πππ¦ ππ‘βππ ππ£πππ‘ ππππ’ππ ]
Median given by middle term; if even number of terms, average of 2 middle terms.
Same applies for Q1, Q2 and IQR.
It is the value of X such that π π β€ π₯ β₯1
2 πππ π π β₯ π₯ β₯
1
2
The mode is the value of x with largest π π = π₯ which can be different from the expected value
ALWAYS watch out for conditional probability (i.e. P(x) given that); it is often implied and not stated
πΈ π = π = π₯ π π = π₯ β‘ πππ₯π =
πππ₯π
ππ
πΌπ π ππ ππππ π‘πππ‘, π‘βππ πΈ ππ = ππΈ π
πΌπ π πππ π πππ ππππ π‘πππ‘π , π‘βππ πΈ ππ + π = ππΈ π + π
πΈ π + π = πΈ π + πΈ π
πΈ[ π π ] = π π₯ π π = π₯
πππ π = π2 = πΈ π β π 2 =π
π π₯π β π 2
ππ
π2 = πΈ π2 β π2
πππ ππ + π = π2πππ π
πππ[π + π] = πππ π + πππ π true only for independent
7
Continuous Random Variables X defined on a β€ x β€ b
probability density function (p.d.f.), f (x), describes the relative likelihood for this variable to take on a given value cumulative distribution function c.d.f.), πΉ π‘ , is found by integrating the p.d.f. from minimum
value of X to t
πΉ π‘ = π π β€ π‘ = β« π π₯ ππ₯π‘
π
For a function π π₯ to be probability function, it must satisfy the following conditions:
1. π π₯ β₯ 0 πππ πππ π₯ π π, π
2. β« π π₯ = 1π
π
3. πππ πππ¦ π β€ π < π β€ π, π π < π < π = β« π π₯ ππ₯π
π
βͺ πΉππ π ππππ‘πππ’ππ’π ππππππ π£πππππππ, π‘βπ ππππππππππ‘π¦ ππ πππ¦ π πππππ π£πππ’π ππ π§πππ
π π = π = 0 β π π β€ π β€ π = π π < π < π = π π β€ π < π ππ‘π.
Median a number m such that β« π π₯ = 1/2π
π
Mode: max on π π , π < π₯ < π (which may not be unique).
ALWAYS watch out for conditional probability (i.e. P(x) given that); it is often implied and not stated
πΈ π = π = β« π₯ π π₯ ππ₯ πΌπ π ππ ππππ π‘πππ‘, π‘βππ πΈ ππ = ππΈ π
πΌπ π πππ π πππ ππππ π‘πππ‘π , π‘βππ πΈ ππ + π = ππΈ π + π
πΈ π + π = πΈ π + πΈ π
πΈ[ π π ] = π π₯ π π = π₯
πππ π = π2 = β« π₯2 π π₯ ππ₯ β [β« π₯π π₯ ππ₯π
π]2π
π
π2 = πΈ π2 β π2
πππ ππ + π = π2πππ π
πππ[π + π] = πππ π + πππ π true only for independent Standard deviation of X: π = βπππ π
CALCULATOR
Binomial Distribution β’ πΏ ~ π©(π, π)
β’ n is number of trials β’ There is either success or failure
β’ p is the probability of a success
β’ (1 β p) is the probability of a failure.
β’ π(π = π₯) = (ππ₯
) ππ₯(1 β π)πβπ₯ π₯ = 0, 1, β¦ , π
β’ πΈ(π) = π = ππ
β’ πππ(π) = π2 = ππ(1 β π)
In a given problem you write: π ~ π΅(100, 0.5) π(π₯ β€ 52) = 0.6913502844
Poisson Distribution β’ πΏ ~ π·π(π)
The average/mean number of occurrences (m) is constant for every interval. The probability of more than one occurrence in a given interval is very small. The number of occurrences in disjoint intervals are independent of each other.
β’ π(π = π₯) =ππ₯πβπ
π₯!, π₯ = 0, 1, 2, β¦
β’ πΈ(π) = π
β’ πππ(π) = π
In a given problem you write: π ~ ππ(0.325) π(π₯ β₯ 6) = 1 β π(π₯ β€ 5) π(π₯ β₯ 6) = .3840393444
If π ~ ππ(β), and π ~ ππ(π), π + π~ ππ(β + π)
BinomCDF(trials , probability of event , value)
o Gives cumulative probability, i.e. the number of successes within n trials is at most the value
ππ‘ πππ π‘: π(π β€ π₯) = ππππππππ(π, π, π₯)
ππ‘ ππππ π‘:
π(π β₯ π₯) = 1 β ππππππππ (π, π, π₯ β 1).
PoissonCDF(mean , value) o Gives cumulative probability, i.e. probability of at most (value) occurrences within a time period
ππ‘ πππ π‘: π(π β€ π₯) = ππππ π πππππ(π, π₯) ππ‘ ππππ π‘:
π(π β₯ π₯) = 1 β ππππ π πππππ (π, π₯ β 1).
BinomPDF(trials , probability of event , value)
o Gives the probability for a particular number of success in n trials
ππ₯πππ‘ππ¦: π(π = π₯ ) = ππππππππ (π, π, π₯)
PoissonPDF(mean , value) o Gives probability of a particular number of occurrences within a time period
ππ₯πππ‘ππ¦: π(π = π₯ ) = ππππ π πππππ (π, π₯)
Normal distribution. β’ πΏ ~ π΅(π, ππ)
β’ π(π₯) =1
πβ2π πβ
12
(π₯βπ
π)
2
β β < π₯ < β
β’ β« π(π₯)ππ₯ = 1β
ββ
β’ π(π₯ β€ π₯1) = ππππ = β« π(π₯)ππ₯π₯1
ββ
β’ π(π₯ β₯ π₯2) = ππππ = β« π(π₯)ππ₯β
π₯2
β’ π(π₯1 β€ π₯ β€ π₯2) = ππππ = β« π(π₯)ππ₯π₯2
π₯1
β’ π(π = π₯1) = 0 β π(π₯1 β€ π β€ π₯2) = π(π₯1 < π < π₯2) = π(π₯1 β€ π < π₯2) ππ‘π.
In a given problem you write: π = 70 SD = 4.5 π ~ π(70, 20.25) π(65 β€ π β€ 80) = 0.8536055925
probability is 85.4%.
In a given problem you write: π = 2870; π = 900 π(π₯ β€ π) = 0.3409
π = 2500
Standardized normal distribution β’ π ~ π΅(π, π)
π =π β π
π (z β score ) π = π + ππ ππ‘ π‘ππππ π’π π€βπππ ππ π‘βπ π£πππ’π ππ π ππ πππππ‘ππππ ππ π
invNorm(probability) o Given a probability, gives the corresponding z-score.
π(π§ β€ π) = Π€(π)
π = πΌππ£ππππ(Π€(π))
π(π β€ π) = π(π < π) ππ π(π = π) = 0
NormalCDF(lower , upper , mean , SD) o Gives probability that a value is within a given range
o percentage of area under a continuous distribution curve from lower bound to upper bound.
π(π₯1 β€ π β€ π₯2) = πππππππππ ( π₯1, π₯2, π, π)
π(π₯ β€ π₯) = πππππππππ (β1πΈ99, π₯, π, π)
π(π β₯ π₯ ) = πππππππππ ( π₯, 1πΈ99, π, π)
InvNorm(probability, π, π) o Given the probability, this function returns the x β value region to the left of x β value.
π(π₯ β€ π) = Π€(π) = ππππ (ππππππππππ‘π¦)
π = πΌππ£ππππ(ππππππππππ‘π¦, π, π)
NormalCDF(lower , upper) o Gives probability that a value is within a given range
π(π§1 β€ π β€ π§2) = πππππππππ ( π§1, π§2)
π(π β€ π§ ) = πππππππππ (β1πΈ99, π§)
π(π β₯ π§ ) = πππππππππ (π§, 1πΈ99)
8