© Copyright Quantopian Inc.
© Modifications Copyright QuantRocket LLC
Licensed under the Creative Commons Attribution 4.0.
By Beha Abasi, Maxwell Margenot, and Delaney Mackenzie
Fundamental data refers to the metrics and ratios measuring the financial characteristics of companies derived from the public filings made by these companies, such as their income statements and balance sheets. Examples of factors drawn from these documents include market cap, net income growth, and cash flow.
This fundamental data can be used in many ways, one of which is to build a linear factor model. Given a set of $k$ fundamental factors, we can represent the returns of an asset, $R_t$, as follows:
$$R_t = \alpha_t + \beta_{t, F_1}F_1 + \beta_{t, F_2}F_2 + ... + \beta_{t, F_k}F_k + \epsilon_t$$where each $F_j$ represents a fundamental factor return stream. These return streams are from portfolios whose value is derived from its respective factor.
Fundamental factor models try to determine characteristics that affect an asset's risk and return. The most difficult part of this is determining which factors to use. Much research has been done on determining significant factors, and what makes things even more difficult is that the discovery of a significant factor often leads to its advantage being arbitraged away! This is one of the reasons why fundamental factor models, and linear factor models in general, are so prevalent in modern finance. Once you have found significant factors, you need to calculate the exposure an asset's return stream has to each factor. This is similar to the calculation of risk premia discussed in the CAPM lecture.
In using fundamental data, we run into the problem of having factors that may not be easily compared due to their varying units and magnitudes. To resolve this, we take two different approaches to bring the data onto the same level - portfolio construction to compare return streams and normalization of factor values.
The first approach consists of using the fundamental data as a ranking scheme and creating a long-short equity portfolio based on each factor. We then use the return streams associated with each portfolio as our model factors.
One of the most well-known examples of this approach is the Fama-French model. The Fama-French model, and later the Carhart four factor model, adds market cap, book-to-price ratios, and momentum to the original CAPM, which only included market risk.
Historically, certain groups of stocks were seen as outperforming the market, namely those with small market caps, high book-to-price ratios, and those that had previously done well (i.e., they had momentum). Empirically, Fama & French found that the returns of these particular types of stocks tended to be better than what was predicted by the security market line of the CAPM.
In order to capture these phenomena, we will use those factors to create a ranking scheme that will be used in the creation of long short equity portfolios. The factors will be $SMB$, measuring the excess return of small market cap companies minus big, $HML$, measuring the excess return of companies with high book-to-price ratios versus low, $MOM$, measuring the excess returns of last month's winners versus last month's losers, and $EXMRKT$ which is a measure of the market risk.
In general, this approach can be used as an asset pricing model or to hedge our portfolios. The latter uses Fama-Macbeth regressions to calculate risk premia, as demonstrated in the CAPM lecture. Hedging can be achieved through a linear regression of portfolio returns on the returns from the long-short factor portfolios. Below are examples of both.
First we import the relevant libraries.
import pandas as pd
import numpy as np
from zipline.pipeline import Pipeline
from zipline.pipeline.data import sharadar
from zipline.pipeline.data import master
from zipline.pipeline.data import EquityPricing
from zipline.pipeline.factors import CustomFactor, Returns, AverageDollarVolume
from zipline.pipeline.filters import AllPresent, All
from zipline.pipeline.classifiers import Classifier
from zipline.research import run_pipeline
import matplotlib.pyplot as plt
Use pipeline to get all of our factor data that we will use in the rest of the lecture.
class Momentum(CustomFactor):
# will give us the returns from last month
inputs = [Returns(window_length=20)]
window_length = 20
def compute(self, today, assets, out, lag_returns):
out[:] = lag_returns[0]
Fundamentals = sharadar.Fundamentals.slice(dimension='ARQ', period_offset=0)
def make_pipeline():
# define our fundamental factor pipeline
pipe = Pipeline()
# market cap and book-to-price data gets fed in here
market_cap = Fundamentals.MARKETCAP.latest
book_to_price = 1/Fundamentals.PB.latest
# and momentum as lagged returns (1 month lag)
momentum = Momentum()
# we also get daily returns
returns = Returns(window_length=2)
TradableStocksUS = (
# Market cap over $500M
(sharadar.Fundamentals.slice(dimension='ARQ', period_offset=0).MARKETCAP.latest >= 500e6)
# dollar volume over $2.5M over trailing 200 days
& (AverageDollarVolume(window_length=200) >= 2.5e6)
# price > $5
& (EquityPricing.close.latest > 5)
# no missing data for 200 days (exclude trading halts, IPOs, etc.)
& AllPresent(inputs=[EquityPricing.close], window_length=200)
& All([EquityPricing.volume.latest > 0], window_length=200)
# common stocks only
& master.SecuritiesMaster.usstock_SecurityType2.latest.eq("Common Stock")
# primary share only
& master.SecuritiesMaster.usstock_PrimaryShareSid.latest.isnull()
)
# we compute a daily rank of both factors, this is used in the next step,
# which is computing portfolio membership
market_cap_rank = market_cap.rank(mask=TradableStocksUS)
book_to_price_rank = book_to_price.rank(mask=TradableStocksUS)
momentum_rank = momentum.rank(mask=TradableStocksUS)
# Grab the top and bottom 1000 for each factor
biggest = market_cap_rank.top(1000)
smallest = market_cap_rank.bottom(1000)
highpb = book_to_price_rank.top(1000)
lowpb = book_to_price_rank.bottom(1000)
top = momentum_rank.top(1000)
bottom = momentum_rank.bottom(1000)
# Define our universe, screening out anything that isn't in the top or bottom
universe = TradableStocksUS & (biggest | smallest | highpb | lowpb | top | bottom)
pipe = Pipeline(
columns = {
'market_cap':market_cap,
'book_to_price':book_to_price,
'momentum':momentum,
'Returns':returns,
'market_cap_rank':market_cap_rank,
'book_to_price_rank':book_to_price_rank,
'momentum_rank':momentum_rank,
'biggest':biggest,
'smallest':smallest,
'highpb':highpb,
'lowpb':lowpb,
'top':top,
'bottom':bottom
},
screen=universe
)
return pipe
# Initializing the pipe
pipe = make_pipeline()
# Now let's start the pipeline
start_date, end_date = '2016-01-01', '2016-12-31'
results = run_pipeline(pipe, start_date=start_date, end_date=end_date, bundle='usstock-1d-bundle')
results.head()
market_cap | book_to_price | momentum | Returns | market_cap_rank | book_to_price_rank | momentum_rank | biggest | smallest | highpb | lowpb | top | bottom | ||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2016-01-04 00:00:00+00:00 | Equity(FIBBG000C2V3D6 [A]) | 1.356254e+10 | 0.307220 | 0.048303 | -0.005826 | 1598.0 | 835.0 | 1630.0 | True | False | False | True | True | False |
Equity(FIBBG006T1NZ18 [AAC]) | 5.281403e+08 | 0.264550 | -0.064742 | 0.027493 | 17.0 | 709.0 | 381.0 | False | True | False | True | False | True | |
Equity(FIBBG005P7Q881 [AAL]) | 2.878697e+10 | 0.130685 | -0.051096 | -0.010514 | 1750.0 | 285.0 | 499.0 | True | False | False | True | False | True | |
Equity(FIBBG000D9V7T4 [AAN]) | 1.803203e+09 | 0.745156 | -0.056796 | -0.007095 | 689.0 | 1600.0 | 459.0 | False | True | True | False | False | True | |
Equity(FIBBG000C2LZP3 [AAON]) | 1.173274e+09 | 0.169062 | -0.022775 | -0.019011 | 421.0 | 396.0 | 803.0 | False | True | False | True | False | True |
Now we can go through the data and build the factor portfolios we want
from quantrocket.master import get_securities
from quantrocket import get_prices
# group_by(level=0).mean() gives you the average return of each day for a particular group of stocks
R_biggest = results[results.biggest]['Returns'].groupby(level=0).mean()
R_smallest = results[results.smallest]['Returns'].groupby(level=0).mean()
R_highpb = results[results.highpb]['Returns'].groupby(level=0).mean()
R_lowpb = results[results.lowpb]['Returns'].groupby(level=0).mean()
R_top = results[results.top]['Returns'].groupby(level=0).mean()
R_bottom = results[results.bottom]['Returns'].groupby(level=0).mean()
# risk-free proxy
securities = get_securities(symbols=['BIL', 'SPY'], vendors='usstock')
BIL = securities[securities.Symbol=='BIL'].index[0]
SPY = securities[securities.Symbol=='SPY'].index[0]
R_F = get_prices('usstock-1d-bundle', sids=BIL, data_frequency='daily', fields='Close', start_date=start_date, end_date=end_date)
R_F = R_F.loc['Close'][BIL].pct_change()[1:]
# find it's beta against market
M = get_prices('usstock-1d-bundle', sids=SPY, data_frequency='daily', fields='Close', start_date=start_date, end_date=end_date)
M = M.loc['Close'][SPY].pct_change()[1:]
# Defining our final factors
EXMRKT = M - R_F
SMB = R_smallest - R_biggest # small minus big
HML = R_highpb - R_lowpb # high minus low
MOM = R_top - R_bottom # momentum
Now that we've constructed our portfolios, let's look at our performance if we were to hold each one.
plt.plot(SMB.index, SMB.values)
plt.ylabel('Daily Percent Return')
plt.legend(['SMB Portfolio Returns']);
plt.plot(HML.index, HML.values)
plt.ylabel('Daily Percent Return')
plt.legend(['HML Portfolio Returns']);
plt.plot(MOM.index, MOM.values)
plt.ylabel('Daily Percent Return')
plt.legend(['MOM Portfolio Returns']);
Now, as we did in the CAPM lecture, we'll calculate the risk premia on each of these factors using the Fama-Macbeth regressions.
import itertools
import statsmodels.api as sm
from statsmodels import regression,stats
import scipy
Our asset returns data is asset and date specific, whereas our factor portfolio returns are only date specific. Therefore, we'll need to spread each day's portfolio return across all the assets for which we have data for on that day.
data = results[['Returns']].set_index(results.index).tz_convert(None, level=0)
asset_list_sizes = [group[1].size for group in data.groupby(level=0)]
# Spreading the factor portfolio data across all assets for each day
SMB_column = [[SMB.tz_convert(None).loc[group[0]]] * size for group, size \
in zip(data.groupby(level=0), asset_list_sizes)]
data['SMB'] = list(itertools.chain(*SMB_column))
HML_column = [[HML.tz_convert(None).loc[group[0]]] * size for group, size \
in zip(data.groupby(level=0), asset_list_sizes)]
data['HML'] = list(itertools.chain(*HML_column))
MOM_column = [[MOM.tz_convert(None).loc[group[0]]] * size for group, size \
in zip(data.groupby(level=0), asset_list_sizes)]
data['MOM'] = list(itertools.chain(*MOM_column))
EXMRKT_column = [[EXMRKT.loc[group[0]]]*size if group[0] in EXMRKT.index else [None]*size \
for group, size in zip(data.groupby(level=0), asset_list_sizes)]
data['EXMRKT'] = list(itertools.chain(*EXMRKT_column))
data = sm.add_constant(data.dropna())
# Our list of assets from pipeline
assets = data.index.get_level_values(1).unique()
# gathering our data to be asset-specific
Y = [data.xs(asset, level=1)['Returns'] for asset in assets]
X = [data.xs(asset, level=1)[['EXMRKT','SMB', 'HML', 'MOM', 'const']] for asset in assets]
# First regression step: estimating the betas
reg_results = [regression.linear_model.OLS(y, x).fit().params \
for y, x in zip(Y, X) if not(x.empty or y.empty)]
indices = [asset for y, x, asset in zip(Y, X, assets) if not(x.empty or y.empty)]
betas = pd.DataFrame(reg_results, index=indices)
betas = sm.add_constant(betas.drop('const', axis=1))
betas = betas.sort_index()
R = data['Returns'].mean(axis=0, level=1).sort_index()
# Second regression step: estimating the risk premia
risk_free_rate = np.mean(R_F)
final_results = regression.linear_model.OLS(R - risk_free_rate, betas).fit()
final_results.summary()
Dep. Variable: | Returns | R-squared: | 0.109 |
---|---|---|---|
Model: | OLS | Adj. R-squared: | 0.107 |
Method: | Least Squares | F-statistic: | 65.24 |
Date: | Tue, 25 May 2021 | Prob (F-statistic): | 3.97e-52 |
Time: | 18:07:10 | Log-Likelihood: | 9317.1 |
No. Observations: | 2146 | AIC: | -1.862e+04 |
Df Residuals: | 2141 | BIC: | -1.860e+04 |
Df Model: | 4 | ||
Covariance Type: | nonrobust |
coef | std err | t | P>|t| | [0.025 | 0.975] | |
---|---|---|---|---|---|---|
const | 0.0011 | 0.000 | 10.204 | 0.000 | 0.001 | 0.001 |
EXMRKT | -4.341e-05 | 0.000 | -0.275 | 0.783 | -0.000 | 0.000 |
SMB | -0.0002 | 5.79e-05 | -3.192 | 0.001 | -0.000 | -7.12e-05 |
HML | 0.0005 | 4.41e-05 | 11.482 | 0.000 | 0.000 | 0.001 |
MOM | 0.0005 | 6.55e-05 | 7.836 | 0.000 | 0.000 | 0.001 |
Omnibus: | 2137.285 | Durbin-Watson: | 2.050 |
---|---|---|---|
Prob(Omnibus): | 0.000 | Jarque-Bera (JB): | 1101421.450 |
Skew: | -3.945 | Prob(JB): | 0.00 |
Kurtosis: | 113.705 | Cond. No. | 5.39 |
As discussed in the CAPM lecture, factor modeling can be used to predict future returns based on current fundamental factors. As well, it could be used to determine when an asset may be mispriced in order to arbitrage the difference, as shown in the CAPM lecture.
Modeling future returns is accomplished by offsetting the returns in the regression, so that rather than predict for current returns, you are predicting for future returns. Once you have a predictive model, the most canonical way to create a strategy is to attempt a long-short equity approach.
Once we've determined that we are exposed to a factor, we may want to avoid depending on the performance of that factor by taking out a hedge. This is discussed in more detail in the Beta Hedging and Risk Factor Exposure lectures. The essential idea is to take the exposure your return stream has to a factor, and short the proportional value. So, if your total portfolio value was $V$, and the exposure you calculated to a certain factor return stream was $\beta$, you would short $\beta V$ amount of that return stream.
The following is an example using the Fama-French factors we used before.
# we'll take a random sample of 500 assets from TradableStocksUS in order to build a random portfolio
random_assets = list(np.random.choice(assets, size=500, replace=False))
portfolio_data = data[data.index.isin(random_assets, level=1)]
# this is the return of our portfolio with no hedging
R_portfolio_time_series = portfolio_data['Returns'].mean(level=0)
# next, we calculate the exposure of our portfolio to each of the Fama-French factors
portfolio_exposure = regression.linear_model.OLS(portfolio_data['Returns'], \
portfolio_data[['EXMRKT', 'SMB', 'HML', 'MOM', 'const']]).fit()
print(portfolio_exposure.summary())
# our hedged return stream
hedged_portfolio = R_portfolio_time_series - \
portfolio_exposure.params[0]*EXMRKT - \
portfolio_exposure.params[1]*SMB.tz_localize(None) - \
portfolio_exposure.params[2]*HML.tz_localize(None) - \
portfolio_exposure.params[3]*MOM.tz_localize(None)
print('Mean, Std of Hedged Portfolio:', np.mean(hedged_portfolio), np.std(hedged_portfolio))
print('Mean, Std of Unhedged Portfolio:', np.mean(R_portfolio_time_series), np.std(R_portfolio_time_series))
OLS Regression Results ============================================================================== Dep. Variable: Returns R-squared: 0.085 Model: OLS Adj. R-squared: 0.085 Method: Least Squares F-statistic: 2527. Date: Tue, 25 May 2021 Prob (F-statistic): 0.00 Time: 18:07:10 Log-Likelihood: 2.5360e+05 No. Observations: 109439 AIC: -5.072e+05 Df Residuals: 109434 BIC: -5.071e+05 Df Model: 4 Covariance Type: nonrobust ============================================================================== coef std err t P>|t| [0.025 0.975] ------------------------------------------------------------------------------ EXMRKT -0.1155 0.009 -13.102 0.000 -0.133 -0.098 SMB 1.4571 0.017 85.475 0.000 1.424 1.491 HML 0.2299 0.019 12.225 0.000 0.193 0.267 MOM -0.5868 0.018 -31.941 0.000 -0.623 -0.551 const 0.0004 7.28e-05 5.418 0.000 0.000 0.001 ============================================================================== Omnibus: 106629.447 Durbin-Watson: 1.761 Prob(Omnibus): 0.000 Jarque-Bera (JB): 172794953.767 Skew: 3.625 Prob(JB): 0.00 Kurtosis: 197.529 Cond. No. 273. ============================================================================== Notes: [1] Standard Errors assume that the covariance matrix of the errors is correctly specified. Mean, Std of Hedged Portfolio: 0.0004005235374867958 0.008435259636217485 Mean, Std of Unhedged Portfolio: 0.000875903244034165 0.01112809172152582
Let's look at a graph of our two portfolio return streams
plt.plot(hedged_portfolio)
plt.plot(R_portfolio_time_series)
plt.legend(['Hedged', 'Unhedged']);
We'll check for normality, homoskedasticity, and autocorrelation in this model. For more information on the tests below, check out the Violations of Regression Models lecture.
For normality, we'll run a Jarque-Bera test, which tests whether our data's skew/kurtosis matches that of a normal distribution. As the standard, we'll reject the null hypothesis that our data is normally distributed if our p-value falls under our confidence level of 5%.
To test for heteroskedasticity, we'll run a Breush-Pagan test, which tests whether the variance of the errors in a linear regression is related to the values of the independent variables. In this case, our null hypothesis is that the data is homoskedastic.
Autocorrelation is tested for using the Durbin-Watson statistic, which looks at the lagged relationship between the errors in a regression. This will give you a number between 0 and 4, with 2 meaning no autocorrelation.
# testing for normality: jarque-bera
_, pvalue_JB, _, _ = stats.stattools.jarque_bera(portfolio_exposure.resid)
print("Jarque-Bera p-value:", pvalue_JB)
# testing for homoskedasticity: breush pagan
_, pvalue_BP, _, _ = stats.diagnostic.het_breuschpagan(portfolio_exposure.resid, \
portfolio_data[['EXMRKT', 'SMB', 'HML', 'MOM', 'const']])
print("Breush Pagan p-value:", pvalue_BP)
# testing for autocorrelation
dw = stats.stattools.durbin_watson(portfolio_exposure.resid)
print("Durbin Watson statistic:", dw)
Jarque-Bera p-value: 0.0 Breush Pagan p-value: 0.03966436711577506 Durbin Watson statistic: 1.7609576485159022
Based on the Jarque-Bera p-value, we would reject the null hypothesis that the data is normally distributed. This means there is strong evidence that our data follows some other distribution.
The test for homoskedasticity suggests that the data is heteroskedastic. However, we need to be careful about this test as we saw that our data may not be normally distributed.
Finally, the Durbin-Watson statistic can be evaluated by looking at the critical values of the statistic. At a confidence level of 95% and 4 explanatory variables, we cannot reject the null hypothesis of no autocorrelation.
Another approach is to normalize factor values for each day and see how predictive of that day's returns they were. This is also known as cross-sectional factor analysis. We do this by computing a normalized factor value $b_{a,j}$ for each asset $a$ in the following way.
$$b_{a,j} = \frac{F_{a,j} - \mu_{F_j}}{\sigma_{F_j}}$$$F_{a,j}$ is the value of factor $j$ for asset $a$ during this time, $\mu_{F_j}$ is the mean factor value across all assets, and $\sigma_{F_j}$ is the standard deviation of factor values over all assets. Notice that we are just computing a z-score to make asset specific factor values comparable across different factors.
The exceptions to this formula are indicator variables, which are set to 1 for true and 0 for false. One example is industry membership: the coefficient tells us whether the asset belongs to the industry or not.
After we calculate all of the normalized scores during time $t$, we can estimate factor $j$'s returns $F_{j,t}$, using a cross-sectional regression (i.e. at each time step, we perform a regression using the equations for all of the assets). Specifically, once we have returns for each asset $R_{a,t}$, and normalized factor coefficients $b_{a,j}$, we construct the following model and estimate the $F_j$s and $a_t$
$$R_{a,t} = a_t + b_{a,F_1}F_1 + b_{a, F_2}F_2 + \dots + b_{a, F_K}F_K$$You can think of this as slicing through the other direction from the first analysis, as now the factor returns are unknowns to be solved for, whereas originally the coefficients were the unknowns. Another way to think about it is that you're determining how predictive of returns the factor was on that day, and therefore how much return you could have squeezed out of that factor.
Following this procedure, we'll get the cross-sectional returns on 2016-11-22, and compute the coefficients for all assets:
We can take the fundamental data we got from the pipeline call above.
date = '2016-11-22'
BTP = results['book_to_price'][date]
z_score = (BTP - BTP.mean()) / BTP.std()
z_score.dropna(inplace=True)
plt.hist(z_score)
plt.xlabel('Z-Score')
plt.ylabel('Frequency');
Notice how there are big outliers in the dataset that cause the z-scores to lose a lot of information. Basically the presence of some very large outliers causes the rest of the data to occupy a relatively small area. We can get around this issue using some data cleaning techniques, such as winsorization.
Winzorization takes the top $n\%$ of a dataset and sets it all equal to the least extreme value in the top $n\%$. For example, if your dataset ranged from 0-10, plus a few crazy outliers, those outliers would be set to 0 or 10 depending on their direction. The following is an example.
# Get some random data
random_data = np.random.normal(0, 1, 100)
# Put in some outliers
random_data[0] = 1000
random_data[1] = -1000
# Perform winsorization
print('Before winsorization', np.min(random_data), np.max(random_data))
scipy.stats.mstats.winsorize(random_data, inplace=True, limits=0.01)
print('After winsorization', np.min(random_data), np.max(random_data))
Before winsorization -1000.0 1000.0 After winsorization -2.0492752290474567 2.5422251610035627
We'll apply the same technique to our data and grab the returns to all the assets in our universe. Then we'll run a linear regression to estimate $F_j$.
BTP = scipy.stats.mstats.winsorize(results['book_to_price'][date], limits=0.01)
BTP_z = (BTP - np.mean(BTP)) / np.std(BTP)
MC = scipy.stats.mstats.winsorize(results['market_cap'][date], limits=0.01)
MC_z = (MC - np.mean(MC)) / np.std(MC)
Lag_Ret = scipy.stats.mstats.winsorize(results['momentum'][date], limits=0.01)
Lag_Ret_z = (Lag_Ret - np.mean(Lag_Ret)) / np.std(Lag_Ret)
returns = results['Returns'][date]
df_day = pd.DataFrame({'R': returns,
'BTP_z': BTP_z,
'MC_z': MC_z,
'Lag_Ret_z': Lag_Ret_z,
'Constant': 1}).dropna()
cross_sectional_results = \
regression.linear_model.OLS(df_day['R'], df_day[['BTP_z', 'MC_z', 'Lag_Ret_z', 'Constant']]).fit()
cross_sectional_results.summary()
Dep. Variable: | R | R-squared: | 0.006 |
---|---|---|---|
Model: | OLS | Adj. R-squared: | 0.004 |
Method: | Least Squares | F-statistic: | 3.685 |
Date: | Tue, 25 May 2021 | Prob (F-statistic): | 0.0116 |
Time: | 18:07:12 | Log-Likelihood: | 4958.5 |
No. Observations: | 1878 | AIC: | -9909. |
Df Residuals: | 1874 | BIC: | -9887. |
Df Model: | 3 | ||
Covariance Type: | nonrobust |
coef | std err | t | P>|t| | [0.025 | 0.975] | |
---|---|---|---|---|---|---|
BTP_z | 0.0003 | 0.000 | 0.625 | 0.532 | -0.001 | 0.001 |
MC_z | 0.0002 | 0.000 | 0.395 | 0.693 | -0.001 | 0.001 |
Lag_Ret_z | 0.0013 | 0.000 | 3.115 | 0.002 | 0.000 | 0.002 |
Constant | 0.0072 | 0.000 | 18.091 | 0.000 | 0.006 | 0.008 |
Omnibus: | 598.658 | Durbin-Watson: | 1.973 |
---|---|---|---|
Prob(Omnibus): | 0.000 | Jarque-Bera (JB): | 14440.986 |
Skew: | 0.935 | Prob(JB): | 0.00 |
Kurtosis: | 16.456 | Cond. No. | 1.16 |
To expand this analysis, you would simply loop through days, running this every day and getting an estimated factor return.
This presentation is for informational purposes only and does not constitute an offer to sell, a solicitation to buy, or a recommendation for any security; nor does it constitute an offer to provide investment advisory or other services by Quantopian, Inc. ("Quantopian") or QuantRocket LLC ("QuantRocket"). Nothing contained herein constitutes investment advice or offers any opinion with respect to the suitability of any security, and any views expressed herein should not be taken as advice to buy, sell, or hold any security or as an endorsement of any security or company. In preparing the information contained herein, neither Quantopian nor QuantRocket has taken into account the investment needs, objectives, and financial circumstances of any particular investor. Any views expressed and data illustrated herein were prepared based upon information believed to be reliable at the time of publication. Neither Quantopian nor QuantRocket makes any guarantees as to their accuracy or completeness. All information is subject to change and may quickly become unreliable for various reasons, including changes in market conditions or economic circumstances.