Quantitative Methods
Study Guide
Rates and Returns
Holding Period Return (HPR) is the total return on an asset over a specific period. Where is the ending price, is the beginning price, and is the cash flow received.
Money-Weighted Rate of Return (MWRR) is the internal rate of return (IRR) on a portfolio, accounting for the timing and amount of all cash flows. It is the rate that sets the present value of inflows equal to the present value of outflows. Best used when the manager controls the timing of cash flows.
Time-Weighted Rate of Return (TWRR) measures the compound growth rate of a portfolio. It removes the effects of cash flow timing and is the standard for investment performance reporting (GIPS). It is calculated by finding the HPR for each sub-period and linking them geometrically.
Annualizing Returns:
- Effective Annual Rate (EAR): Accounts for compounding within a year.
- To convert a holding period return to an EAR: , where t is the number of days in the holding period.
Time Value of Money in Finance
The core principle is that money available now is worth more than the same amount in the future due to its potential earning capacity.
Key Components:
- PV: Present Value
- FV: Future Value
- I/Y: Interest Rate per period
- N: Number of compounding periods
- PMT: Annuity payment
Future Value (FV) of a Single Sum:
Present Value (PV) of a Single Sum:
Annuities: A series of equal cash flows at regular intervals.
- Ordinary Annuity: Cash flows occur at the end of each period.
- Annuity Due: Cash flows occur at the beginning of each period. The PV and FV of an annuity due are greater than an ordinary annuity.
Perpetuity: An annuity that continues forever.
Statistical Measures of Asset Returns
Measures of Central Tendency:
- Arithmetic Mean: Simple average. Best for forecasting a single period's return.
- Geometric Mean: Measures compound growth rate over time. Always less than or equal to the arithmetic mean.
- Harmonic Mean: Used for averaging ratios (e.g., dollar-cost averaging).
Measures of Dispersion:
- Variance (): Average of the squared deviations from the mean.
- Population Variance:
- Sample Variance:
- Standard Deviation (): Square root of the variance. Measures the volatility of returns.
- Coefficient of Variation (CV): A measure of relative dispersion. Useful for comparing risk of assets with different expected returns. (Standard Deviation / Mean)
Measures of Shape:
- Skewness: Measures the asymmetry of a distribution.
- Positive Skew: Long tail to the right (Mean > Median > Mode).
- Negative Skew: Long tail to the left (Mean < Median < Mode).
- Kurtosis: Measures the degree to which a distribution is more or less "peaked" than a normal distribution.
- Leptokurtic: More peaked, fatter tails (Kurtosis > 3).
- Platykurtic: Less peaked, thinner tails (Kurtosis < 3).
- Mesokurtic: Normal distribution (Kurtosis = 3).
Probability Trees and Conditional Expectations
Key Concepts:
- Conditional Probability : The probability of event A occurring, given that event B has occurred.
- Total Probability Rule: Used to find the unconditional probability of an event, given conditional probabilities.
- Expected Value : The weighted average of all possible outcomes.
- Covariance: Measures how two variables move together.
- Correlation: A standardized measure of covariance, ranging from -1 to +1.
- Bayes' Theorem: Updates a prior probability with new information to get a posterior probability.
A probability tree is a visual tool to represent outcomes and their associated probabilities in a sequence of events.
Portfolio Mathematics
The risk and return of a portfolio depend on the risk/return of its individual assets and the correlation between them.
Asset | Weight (w) | Expected Return E(R) | Standard Deviation () |
---|---|---|---|
Stock A | |||
Stock B |
Portfolio Expected Return:
Portfolio Variance: where is the correlation coefficient between A and B.
Key Insight: Diversification benefits increase as the correlation between assets decreases. Perfect negative correlation () can potentially eliminate all risk.
Simulation Methods
Monte Carlo Simulation is a computer-based method that uses random sampling to model the probability of different outcomes in a process that cannot easily be predicted due to the intervention of random variables.
Steps:
- Specify the model: Define the quantity of interest and the variables that influence it.
- Define probability distributions: Specify distributions for the random variables.
- Generate random inputs: Draw random values from the specified distributions.
- Run the simulation: Calculate the quantity of interest thousands of times.
- Analyze the results: The distribution of outcomes provides an estimate of the expected value and risk of the quantity.
Applications: Valuing complex derivatives, modeling portfolio risk (VaR), financial planning. Limitations: Highly dependent on the specified model and input assumptions ("garbage in, garbage out").
Estimation and Inference
Estimation involves using sample data to estimate population parameters (e.g., using sample mean to estimate population mean ).
Central Limit Theorem (CLT): For any population with mean and variance , the sampling distribution of the sample mean for a large sample size (n ≥ 30) will be approximately normal with mean and variance . This is crucial because it allows us to make inferences about the population mean without knowing the population's distribution.
Confidence Intervals: An interval estimate provides a range of values within which the true population parameter is expected to lie, with a specified degree of confidence. Formula: Point Estimate ± (Reliability Factor × Standard Error)
- Point Estimate: The sample statistic (e.g., ).
- Reliability Factor: A number based on the assumed distribution and the confidence level (e.g., 1.96 for 95% confidence using a z-statistic).
- Standard Error: The standard deviation of the sample statistic (e.g., ).
Hypothesis Testing
A formal procedure for accepting or rejecting a statement (hypothesis) about a population parameter based on sample data.
The Process:
- State the hypotheses:
- Null Hypothesis (): The hypothesis to be tested. Usually a statement of "no effect" or "no difference." It contains an equality sign (=, ≤, or ≥).
- Alternative Hypothesis (): The hypothesis accepted if the null is rejected. It contains an inequality sign (≠, <, or >).
- Select the test statistic: e.g., z-test, t-test.
- Specify the level of significance (): The probability of a Type I error (e.g., 5%).
- State the decision rule: Compare the test statistic to a critical value or compare the p-value to .
- Calculate the test statistic from the sample data.
- Make a decision: Reject or fail to reject the null hypothesis.
Decision | is True | is False |
---|---|---|
Do Not Reject | Correct Decision | Type II Error () |
Reject | Type I Error () | Correct Decision (Power = 1-) |
p-value: The smallest level of significance () at which can be rejected.
- If p-value < , reject .
- If p-value ≥ , fail to reject .
Parametric and Non-Parametric Tests of Independence
Tests used to determine if a relationship exists between two variables.
Feature | Parametric Tests | Non-Parametric Tests |
---|---|---|
Assumptions | Require strong assumptions about the population distribution (e.g., normality). | Make few or no assumptions about the population distribution. |
Data Type | Interval or ratio data. | Nominal or ordinal data (ranks). Can also be used for interval/ratio if parametric assumptions are violated. |
Power | More powerful if assumptions are met. | Less powerful than parametric tests if assumptions are met. |
Example | t-test for the significance of a correlation coefficient. | Spearman rank correlation test, Chi-square test of independence. |
- Use a parametric test when the data is known to be normally distributed and other assumptions are met.
- Use a non-parametric test for ranked data or when you cannot verify the assumptions of a parametric test.
Simple Linear Regression
Models the linear relationship between one dependent variable (Y) and one independent variable (X).
Model Equation:
- : Dependent variable
- : Independent variable
- : Intercept (estimated value of Y when X=0)
- : Slope coefficient (estimated change in Y for a one-unit change in X)
- : Error term (the portion of Y not explained by X)
Key Regression Outputs:
- Coefficient of Determination (): The percentage of variation in the dependent variable (Y) that is explained by the independent variable (X). Ranges from 0 to 1.
- Standard Error of the Estimate (SEE): Measures the typical distance between the observed values and the values predicted by the regression line. A smaller SEE indicates a better model fit.
- F-test: Tests the overall significance of the regression model. A significant F-test indicates that at least one independent variable is significant.
- t-test: Tests the significance of individual coefficients ( and ). A significant t-test for indicates a statistically significant relationship between X and Y.
Introduction to Big Data Techniques
Big Data is characterized by its high Volume, Velocity, and Variety (the 3 V's). It requires advanced tools to capture, store, manage, and analyze.
Key Techniques:
- Machine Learning (ML): Algorithms that learn from data to make predictions or decisions.
- Supervised Learning: Uses labeled data (e.g., regression, classification). Application: Predicting credit default.
- Unsupervised Learning: Uses unlabeled data to find patterns (e.g., clustering). Application: Customer segmentation.
- Deep Learning: A subset of ML using neural networks with many layers.
- Natural Language Processing (NLP): Enables computers to understand and process human language. Application: Analyzing sentiment from news articles or social media posts (text mining).
- Data Visualization: The graphical representation of information and data. Application: Using heat maps or tree maps to visualize portfolio exposures.
Challenges in Finance:
- Overfitting: Creating a model that is too complex and fits the training data's noise, leading to poor performance on new data.
- Data Quality: Unstructured and complex data can be noisy and require significant cleaning.
- Spurious Correlation: Finding relationships that are due to chance rather than a true underlying cause.