Anda di halaman 1dari 3

Stats Final Review quantitative vs categorical variables: Categorical variables take on values that are names or labels; averaging

does not make sense. Quantitative variables are numeric. They represent a measurable quantity. Numerical data where averaging makes sense. 5- Number Summary: the smallest observation , first quartile, median, third quartile, and largest observation, written in order from smallest to largest. Mean and Standard Deviation: Mean is the average. Standard Deviation is the square root of the variance, measures spread. Means can be added. Standard deviations cannot be combined, they have to be turned into variances first. Stemplot; back-to-back stemplot:

Stemplot

Back-to-back Stemplot

shape of data: Symmetry, Number of peaks, skewness: Distributions with most of their observations on the left (toward lower values) are said to be skewed right; and distributions with most of their observations on the right (toward higher values) are said to be skewed left. Uniform, when the observations in a set of data are equally spread across the range of the distribution. When is mean>median, median>mean, and mean=median: The mean is greater than the median when the distribution is skewed right. The median is greater than the mean when the distribution is skewed left. The mean is equal to the median in symmetric distributions. Percentile;changing from data to percentile: The values that divide a rank-ordered set of elements into 100 equal parts are called percentiles. 68-95-99.7 rule (Empirical Rule): make modified boxplots; side by side boxplots:

Outliers using 1.5 x IQR: Step 1: Q3- Q1 called IQR Step 2: 1.5 times IQR = x Step 3: Subtract x from Q1; Add x to Q3 Step 4: Any data outside those answers would be an outlier. Differences in mean and median as measures of center: The mean describes the center as an average value, where the actual values of the data points play an important role since we are summing values and then dividing by count (number of values). The median, on the other hand, locates the middle value as the center, and the order of the data is the key to finding it (here we are sorting and only counting). frequency histogram: A graph that uses vertical columns to show frequencies (how many times each score occurs). relative frequency histogram: A relative frequency histogram uses the same information as a frequency histogram but compares each class interval to the total number of items. locate median of relative frequency histogram: computing Z probabilities: Z= X - X/ Standard Deviation LSRL passes through (x, y) Interpretation of various values of r; calculating r: r is between -1 and 1. .6 to 1 is very strong. .3 to .5 is weak. 0 to .3 is very weak. If slope is positive, r is positive, if slope is negative, r is negative. R= r is the percent of the variation between x and y explained by the LSRL. Explanatory/response variables:The response variable is the outcome of a study. A variable you would be interested in predicting or forecasting. Often called a dependent variable or predicted variable. The explanatory variable is any variable that explains the response variable. Often called an independent variable or predictor variable. On graph, response is on y-axis, explanatory is on x-axis. given a relationship(equation), find value at certain x; find residual if given observed value: plug in x into equation, and residual = actual - prediction given a situation, write the equation, tell slope of line: plot given data, make scatterplot, is data likely to be linear or exponential: find LSRL with calculator, plot it with scatterplot points, is it a valid model: computer logs with lists, show transformed data, find LSRL on transformed data: use LSRL to predict non-listed data points: plug in x into LSRL. If x is outside range of data, then the data is risky and therefore extrapolation. designing a study (description and/or diagram); assigning #'s for randomization: reading random digit table to collect data; report results: seeding the calculator to produce same random results as other people: types of samples: SRS (requirements?), stratified, multistage, block, matched pairs, double blind, etc.: In simple random sampling, each member of a population has an equal chance of being included in the sample. Also, each combination of members of the population has an equal chance of composing the sample. difference between study and experiment:The fundamental difference is that an experiment requires you to observe what happens when you make some sort of change. An observational study finds the relationship between changes that already exist. 3 principles of good experimental design: Randomization, replication, and control of lurking variables. probability rules disjoint events, independence, etc: Two events are mutually exclusive if they cannot occur at the same time. Another word that means mutually exclusive is disjoint. If two events are disjoint, then the probability of them both occurring at the same time is 0. Two events are independent if the occurrence of

one does not change the probability of the other occurring. If events are independent, then the probability of them both occurring is the product of the probabilities of each occurring. If the occurrence of one event does affect the probability of the other occurring, then the events are dependent. multiplication rule and addition rules.. requirements...Venn Diagrams.. Tree Diagrams: discrete and continuous random variables: Discrete variables are items that you measure. Continuous variables are items that can be counted. density curves: valid?: probability distributions: Probability distributions are valid if all of its data adds up to 1, and all of the values are greater than or equal to 0. mean,variance, std deviation of random variable: to find mean: find average of all values of P(X) to find variance: square each value of P(X) and add them all together. To find standard deviation: square root of variance Rules for Means; Rules for variances: expected value (expectation): (the probability of winning)(amount you can win)=(fair cost to play) fair games: A game is FAIR if the expected value (expectant) is the same as the cost to play. Law of Large Numbers: This says that the actual mean of many trials gets closer to the distribution mean as more trials are made.

Anda mungkin juga menyukai