33 Event Studies
The event study methodology is widely used in finance, marketing, and management to measure the impact of specific events on stock prices. The foundation of this methodology is the Efficient Markets Hypothesis proposed by Fama (1970), which asserts that asset prices reflect all available information. Under this assumption, stock prices should immediately react to new, unexpected information, making event studies a useful tool for assessing the economic impact of firm- and non-firm-initiated activities.
A note on descriptive vs. causal interpretation. In its classical finance form, an event study produces a series of abnormal returns: deviations of observed returns from the return predicted by a market (or factor) model. Abnormal returns are a descriptive object. They tell us how prices moved relative to a statistical benchmark around the event window.
Interpreting them as a causal effect of the event on firm value requires additional design-level assumptions:
- The event date is known precisely and was not anticipated. Otherwise the market has priced it in before the window opens.
- No other news or events overlap the event window, so that the price movement is attributable to the focal event and not a confounder.
- The expected-return model is correctly specified during the estimation window.
When these conditions are not met, the event study documents a correlation between a period and a price movement, not a causal effect. In modern empirical practice, event studies are often layered on top of an explicit quasi-experimental design (DiD, RD, or IV): the event-study plot then serves as a visualization of dynamic treatment effects, and the causal identification comes from the underlying design rather than from the event study itself.
The first event study was conducted by Dolley (1933), while Campbell et al. (1998) formalized the methodology for modern applications. Later, Dubow and Monteiro (2006) developed a metric to assess market transparency (i.e., a way to gauge how “clean” a market is) by tracking unusual stock price movements before major regulatory announcements. Their study found that abnormal price shifts before announcements could indicate insider trading, as prices reacted to leaked information before official disclosures.
Advantages of Event Studies
More Reliable than Accounting-Based Measures: Unlike financial metrics (e.g., profits), which managers can manipulate, stock prices are harder to alter and reflect real-time investor sentiment (Benston 1985).
Easy to Conduct: Event studies require only stock price data and simple econometric models, making them widely accessible for researchers.
Types of Events in Event Studies
Table 33.1 groups the events typically analysed by event studies into firm-internal and firm-external categories.
| Event Type | Examples |
|---|---|
| Internal Events | Stock repurchase, earnings announcements, leadership changes |
| External Events | Macroeconomic shocks, regulatory changes, media reports |
33.1 A Brief Tour of How the Method Travelled
It is worth pausing on how the event study became a workhorse in three quite different fields, because the migration from finance into management and then marketing tells you something about what the method actually buys you and what it cannot.
The technique was born in finance. Fama et al. (1969) used the announcement of stock splits to ask a deceptively simple question: do prices adjust to new information immediately, or does adjustment take place over days or weeks? The answer mattered for the efficient-markets hypothesis, but the machinery, pick an event, define a window, compute a normal return, attribute the gap to the event, was general. Once the machinery existed, every field with a clean event and a market response found a use for it.
Management researchers were the next adopters. By the late 1990s the question had shifted from market efficiency to managerial decision-making: do the choices managers make actually create shareholder value, and which ones? McWilliams and Siegel (1997) surveyed how the method had been imported into management research and laid out the methodological cautions, short windows, careful event definition, attention to confounding events, that have shaped practice ever since. Their review remains the standard starting point for anyone applying event studies outside finance.
Marketing scholars adopted the same machinery to defend marketing decisions in the language that boards and CFOs already spoke: stock price. If a brand acquisition, a celebrity endorsement, or a new product launch is supposed to create value, the event study lets us check whether the market thought so on the day the news broke. Two strands of marketing event studies emerged. The first asks about firm-initiated events: things the firm chose to do. The second asks about non-firm-initiated events: things that happened to the firm, often outside its control.
Tables 33.2 and 33.3 collect representative work in each strand, organised by event type so that the table doubles as a reading list for a particular research question. The point is not to memorise these citations but to see the breadth of what an event study can address, from corporate name changes to data breaches, once you accept that “value created” is being proxied by “market reaction”.
| Event Type | Studies |
|---|---|
| Corporate Changes | (Horsky and Swyngedouw 1987) (name change), (Kalaignanam and Bahadir 2013) (corporate brand change) |
| New Product Strategies | (Chaney, Devinney, and Winer 1991) (new product announcements), (Raassens, Wuyts, and Geyskens 2012) (outsourcing product development), (Sood and Tellis 2009) (innovation payoff), (Borah and Tellis 2014) (make, buy or ally for innovations), (Fang, Lee, and Yang 2015) (co-development agreements) |
| Brand & Marketing Strategies | (Lane and Jacobson 1995) (brand extensions), (Wiles, Morgan, and Rego 2012) (brand acquisition) |
| Advertising & Promotions | (Wiles et al. 2010) (deceptive advertising), (Cornwell, Pruitt, and Clark 2005) (sponsorship announcements) |
| Strategic Alliances | (Houston and Johnson 2000) (joint ventures), (Fang, Lee, and Yang 2015) (co-development agreements), (A. B. Sorescu, Chandy, and Prabhu 2007) (M&A), (Homburg, Vollmayr, and Hahn 2014) (channel expansions) |
| Entertainment & Celebrity Endorsements | (Agrawal and Kamakura 1995) (celebrity endorsements), (Elberse 2007) (casting announcements), (Wiles and Danielova 2009) (product placement in movies), (Joshi and Hanssens 2009) (movie releases), (Karniouchina, Uslay, and Erenburg 2011) (product placement), (Mazodier and Rezaee 2013) (sports announcements) |
Two further marketing studies sit slightly outside that grid but illustrate the same logic: Geyskens, Gielens, and Dekimpe (2002) examined newspapers’ decisions to launch internet channels, and Boyd, Chandy, and Cunha Jr (2010) looked at the market’s reaction to new CMO appointments. Both treat a strategic choice as the event and ask whether shareholders rewarded or punished it.
The non-firm-initiated literature is structurally similar but methodologically more demanding. Because the firm did not choose the event, the researcher has fewer levers to manipulate: there is no internal documentation of the announcement timing, leakage is harder to rule out, and selection into “events that journalists write about” is itself a research question. The studies in Table 33.3 tackle this in different ways, some focus on regulator-announced events with a clean publication date, others on negative shocks where the firm’s response is partly endogenous to the event itself.
| Event Type | Studies |
|---|---|
| Regulatory Decisions | (A. B. Sorescu, Chandy, and Prabhu 2003) (FDA approvals), (Rao, Chandy, and Prabhu 2008) (FDA approvals), (Tipton, Bharadwaj, and Robertson 2009) (deceptive advertising regulation) |
| Media Coverage & Consumer Reactions | (Jacobson and Mizik 2009) (customer satisfaction score release), (Y. Chen, Liu, and Zhang 2012) (third-party movie reviews), (Tellis and Johnson 2007) (quality reviews by Walter Mossberg) |
| Economic & Market Shocks | (Gielens et al. 2008) (Walmart’s entry into the UK), (Xiong and Bharadwaj 2013) (asymmetric news impact), (Pandey, Shanahan, and Hansen 2005) (diversity elite list) |
| Consumer & Industry Recognitions | (Balasubramanian, Mathur, and Thakur 2005) (high-quality achievements), (Fornell et al. 2006) (customer satisfaction), (Ittner, Larcker, and Taylor 2009) (customer satisfaction) |
| Financial & Market Reactions | (Boyd and Spekman 2008) (indirect ties), (Karniouchina, Moore, and Cooney 2009) (Mad Money with Jim Cramer), (Bhagat, Bizjak, and Coles 1998) (litigation) |
| Product & Service Failures | (Y. Chen, Ganesan, and Liu 2009) (product recalls), (Gao et al. 2015) (product recalls), (Malhotra and Kubowicz Malhotra 2011) (data breach) |
The space of plausible event studies is far from exhausted. Major advertising campaigns, market entries, product recalls, and patent announcements are all candidate events whose informational content arrives with a clean timestamp and whose effect on firm value is in principle measurable. Whenever you encounter a setting where (a) the event is dated precisely, (b) market participants could plausibly update on it, and (c) you can construct a defensible counterfactual from the estimation window, the apparatus that follows in this chapter is in scope.
33.2 Key Assumptions
The interpretation of an event study, whether descriptive or causal, rests on a small set of assumptions about how prices, information, and the firm’s stakeholders interact. None of these are testable in the strong sense that randomization is, and each fails in identifiable ways. Reading the four assumptions below as licensing claims, each one buying you a specific interpretation, is more useful than reading them as a checklist.
The Efficient Market Hypothesis (Fama 1970) is the load-bearing assumption. It says that stock prices fully and instantly reflect publicly available information, so a measurable price reaction in the days surrounding an event is interpretable as the market’s revaluation of the firm in light of the event’s news content. When EMH holds in its strong form, the event window is exactly where the news shows up; when it holds only weakly (information leaks before the announcement, prices drift in the days after), the analyst must widen the window or accept that the estimated abnormal return is contaminated. EMH is not all-or-nothing in practice, and most modern event studies assume a semi-strong form (prices reflect public information) rather than a strong form (prices reflect all information including private signals).
The stock market as a proxy for firm value assumption says that the price of equity is a meaningful summary of the firm’s value to its primary stakeholders. This is uncontroversial in finance applications where shareholders are the relevant audience. It is more contested in marketing or management applications, where the relevant audience is sometimes customers, employees, or regulators, and where firm value to those stakeholders is not perfectly tracked by share prices. The closer the research question is to “what did this event do to shareholder wealth?”, the better this assumption holds; the further it is from that question, the more the analyst should report alternative outcome measures alongside abnormal returns.
The sharp event-effect assumption requires that the event causes an immediate and concentrated price reaction. If the market reacts gradually over several months, no plausible event window will isolate the effect; if the market anticipates the event, the price has already moved before the window opens. The assumption is most credible for events with precise timestamps that the market could not have anticipated (regulatory rulings on a fixed announcement date, court verdicts, corporate disclosures with mandated reporting calendars). It is least credible for events whose probability of occurrence was already partially priced in, in which case the abnormal return measures the surprise component, not the full effect of the event.
The proper calculation of expected returns is what gives the abnormal return a benchmark. The mechanics matter: a market model with the wrong factor specification, an estimation window that includes a structural break, or a beta estimated on a thin sample will all produce expected returns that misstate the counterfactual price path. The downstream consequence is that the abnormal return is only as clean as the expected-return model. We return to this point in the expected return calculation section and the econometric event-study designs section.
These four assumptions together support a descriptive claim, that the abnormal return measures the price reaction to the event window. Promoting that descriptive claim to a causal claim, that the event itself caused the price reaction, requires the additional design-level assumptions discussed at the start of this chapter: no contemporaneous confounding events, no anticipation, and a benchmark model that is correctly specified throughout the estimation window. Modern empirical practice typically pairs the event-study machinery with an explicit quasi-experimental design (Regression Discontinuity, examiner IV, or Difference-in-Differences) so that the causal claim does not rest on the four assumptions above alone.
33.3 Steps for Conducting an Event Study
33.3.1 Step 1: Event Identification
An event study examines how a particular event affects a firm’s stock price, assuming that stock markets incorporate new information efficiently. The event must influence either the firm’s expected cash flows or discount rate (A. Sorescu, Warren, and Ertekin 2017, 191).
Common Types of Events Analyzed
Table 33.4 lists the broad event categories most often analysed and gives concrete examples of each.
| Event Category | Examples |
|---|---|
| Corporate Actions | Dividends, mergers & acquisitions (M&A), stock buybacks, name changes, brand extensions, sponsorships, product launches, advertising campaigns |
| Regulatory Changes | New laws, taxation policies, financial deregulation, trade agreements |
| Market Events | Privatization, nationalization, entry/exit from major indices |
| Marketing-Related Events | Celebrity endorsements, new product announcements, media reviews |
| Crisis & Negative Shocks | Product recalls, data breaches, lawsuits, financial fraud scandals |
To systematically identify events, researchers use WRDS S&P Capital IQ Key Developments, which tracks U.S. and international corporate events.
33.3.2 Step 2: Define the Event and Estimation Windows
33.3.2.1 (A) Estimation Window (\(T_0 \to T_1\))
The estimation window is used to compute normal (expected) returns before the event. Table 33.5 summarises common choices in the literature.
| Study | Estimation Window |
|---|---|
| (Johnston 2007) | 250 days before the event, with a 45-day gap before the event window |
| (Wiles, Morgan, and Rego 2012) | 90-trading-day estimation window ending 6 days before the event |
| (A. Sorescu, Warren, and Ertekin 2017, 194) | 100 days before the event |
Leakage Concern: To avoid biases from information leaking before the event, researchers should check broad news sources (e.g., LexisNexis, Factiva, RavenPack) for pre-event rumors.
33.3.2.2 (B) Event Window (\(T_1 \to T_2\))
The event window captures the market’s reaction to the event. The selection of an appropriate window length depends on event type and information speed; Table 33.6 reports choices used in the literature.
| Study | Event Window |
|---|---|
| (Balasubramanian, Mathur, and Thakur 2005; Boyd, Chandy, and Cunha Jr 2010; Fornell et al. 2006) | 1-day window |
| (Raassens, Wuyts, and Geyskens 2012; Sood and Tellis 2009) | 2-day window |
| (Cornwell, Pruitt, and Clark 2005; A. B. Sorescu, Chandy, and Prabhu 2007) | Up to 10 days |
33.3.3 Step 3: Compute Normal vs. Abnormal Returns
The abnormal return measures how much the stock price deviates from its expected return:
\[ \text{AR}_{it} = R_{it} - E(R_{it} \mid X_t) \]
where:
\(\text{AR}_{it}\) = abnormal return for firm \(i\) at time \(t\)
\(R_{it}\) = realized (dividend-adjusted) return
\(E(R_{it} \mid X_t)\) = expected return given the conditioning information \(X_t\) (typically the market or factor returns over the estimation window)
33.3.3.1 (A) Statistical Models for Expected Returns
These models assume jointly normal and independently distributed returns.
-
Constant Mean Return Model
\[ E(R_{it}) = \frac{1}{T} \sum_{t=T_0}^{T_1} R_{it} \] -
Market Model
\[ R_{it} = \alpha_i + \beta_i R_{mt} + \epsilon_{it} \] -
Adjusted Market Return Model
\[ E(R_{it}) = R_{mt} \]
33.3.4 Step 4: Compute Cumulative Abnormal Returns
Once abnormal returns are computed, we aggregate them over the event window:
\[ CAR_{i} = \sum_{t=T_{\text{event, start}}}^{T_{\text{event, end}}} AR_{it} \]
For multiple firms, compute the Average Cumulative Abnormal Return (ACAR):
\[ ACAR = \frac{1}{N} \sum_{i=1}^{N} CAR_{i} \]
33.4 Event Studies in Marketing
A key challenge in marketing-related event studies is determining the appropriate dependent variable (Skiera, Bayer, and Schöler 2017). Traditional event studies in finance use cumulative abnormal returns (CAR) on shareholder value (\(CAR^{SHV}\)). However, marketing events primarily affect a firm’s operating business, rather than its total shareholder value, leading to potential distortions if financial leverage is ignored.
According to valuation theory, a firm’s shareholder value (\(SHV\)) consists of three components (Schulze, Skiera, and Wiesel 2012):
\[ SHV = \text{Operating Business Value} + \text{Non-Operating Assets} - \text{Debt} \]
Many marketing-related events primarily impact operating business value (e.g., brand perception, customer satisfaction, advertising efficiency), while non-operating assets and debt remain largely unaffected.
Ignoring firm-specific leverage effects in event studies can cause:
- Inflated impact for firms with high debt.
- Deflated impact for firms with large non-operating assets.
Thus, it is recommended that both \(CAR^{OB}\) and \(CAR^{SHV}\) be reported, with justification for which is most appropriate.
Surprisingly few event studies have explicitly controlled for financial structure. Two exceptions are worth flagging because they show how the correction can be operationalised. Chaney, Devinney, and Winer (1991) look at the relationship between advertising expenses and firm value while explicitly holding leverage in the picture, and Gielens et al. (2008) extend the same logic to marketing-spending shocks. Outside this small literature, the implicit assumption is that leverage is uncorrelated with the event of interest, which is rarely defended and often demonstrably false.
33.4.1 Definition
- Cumulative Abnormal Return on Shareholder Value (\(CAR^{SHV}\))
\[ CAR^{SHV} = \frac{\sum \text{Abnormal Returns}}{SHV} \]
-
Shareholder Value (\(SHV\)): Market capitalization, defined as:
\[ SHV = \text{Share Price} \times \text{Shares Outstanding} \]
- Cumulative Abnormal Return on Operating Business (\(CAR^{OB}\))
To correct for leverage effects, \(CAR^{OB}\) is calculated as:
\[ CAR^{OB} = \frac{CAR^{SHV}}{\text{Leverage Effect}} \]
where:
\[ \text{Leverage Effect} = \frac{\text{Operating Business Value}}{\text{Shareholder Value}} \]
Key Relationships:
- Operating Business Value = \(SHV -\) Non-Operating Assets \(+\) Debt.
- Leverage Effect (\(LE\)) measures how a 1% change in operating business value translates into shareholder value movement.
- Leverage Effect vs. Leverage Ratio
Leverage Effect (\(LE\)) is not the same as the leverage ratio, which is typically:
\[ \text{Leverage Ratio} = \frac{\text{Debt}}{\text{Firm Size}} \]
where firm size can be:
Book value of equity
Market capitalization
Total assets
Debt + Equity
33.4.2 When Can Marketing Events Affect Non-Operating Assets or Debt?
While most marketing events impact operating business value, in rare cases they also influence non-operating assets and debt (Table 33.7).
| Marketing Event | Impact on Financial Structure |
|---|---|
| Excess Pre-ordering (G. C. Hall, Hutchinson, and Michaelas 2004) | Affects short-term debt |
| CMO Turnover (Berger, Ofek, and Yermack 1997) | Higher debt due to manager turnover |
| Unique Product Development (Bhaduri 2002) | Alters debt levels |
These exceptions highlight why controlling for financial structure is crucial in event studies.
33.4.3 Calculating the Leverage Effect
We can express leverage effect (\(LE\)) as:
\[ \begin{aligned} LE &= \frac{\text{Operating Business Value}}{\text{Shareholder Value}} \\ &= \frac{(\text{SHV} - \text{Non-Operating Assets} + \text{Debt})}{\text{SHV}} \\ &= \frac{prcc_f \times csho - ivst + dd1 + dltt + pstk}{prcc_f \times csho} \end{aligned} \]
where:
\(prcc_f\) = Share price
\(csho\) = Common shares outstanding
\(ivst\) = Short-term investments (Non-Operating Assets)
\(dd1\) = Long-term debt due in one year
\(dltt\) = Long-term debt
\(pstk\) = Preferred stock
33.4.4 Computing Leverage Effect from Compustat Data
Figure ?? shows the cross-firm distribution of the leverage effect computed from Compustat together with the within-firm coefficient of variation across years.
# Load required libraries
library(tidyverse)
# Load dataset
df_leverage_effect <- read.csv("data/leverage_effect.csv.gz") %>%
# Filter active firms
filter(costat == "A") %>%
# Drop missing values
drop_na() %>%
# Compute Shareholder Value (SHV)
mutate(shv = prcc_f * csho) %>%
# Compute Operating Business Value (OBV)
mutate(obv = shv - ivst + dd1 + dltt + pstk) %>%
# Compute Leverage Effect
mutate(leverage_effect = obv / shv) %>%
# Remove infinite values and non-positive leverage effects
filter(is.finite(leverage_effect), leverage_effect > 0) %>%
# Compute within-firm statistics
group_by(gvkey) %>%
mutate(
within_mean_le = mean(leverage_effect, na.rm = TRUE),
within_sd_le = sd(leverage_effect, na.rm = TRUE)
) %>%
ungroup()
# Summary statistics
mean_le <- mean(df_leverage_effect$leverage_effect, na.rm = TRUE)
max_le <- max(df_leverage_effect$leverage_effect, na.rm = TRUE)
# Plot histogram of leverage effect
hist(
df_leverage_effect$leverage_effect,
main = "Distribution of Leverage Effect",
xlab = "Leverage Effect",
col = "blue",
breaks = 30
)
Figure 33.1: Cross-firm distribution of the Compustat leverage effect (top) and the within-firm coefficient of variation across years (bottom).
# Compute coefficient of variation (CV)
cv_le <-
sd(df_leverage_effect$leverage_effect, na.rm = TRUE) / mean_le * 100
# Plot within-firm coefficient of variation histogram
df_leverage_effect %>%
group_by(gvkey) %>%
slice(1) %>%
ungroup() %>%
mutate(cv = within_sd_le / within_mean_le) %>%
pull(cv) %>%
hist(
main = "Within-Firm Coefficient of Variation",
xlab = "CV",
col = "red",
breaks = 30
)
Figure 33.2: Cross-firm distribution of the Compustat leverage effect (top) and the within-firm coefficient of variation across years (bottom).
33.5 Economic Significance
The total wealth gain (or loss) resulting from a marketing event is given by:
\[ \Delta W_t = CAR_t \times MKTVAL_0 \]
where:
- \(\Delta W_t\) = Change in firm value (gain or loss).
- \(CAR_t\) = Cumulative abnormal return up to date \(t\).
- \(MKTVAL_0\) = Market value of the firm before the event window.
Interpretation:
- If \(\Delta W_t > 0\): The event increased firm value.
- If \(\Delta W_t < 0\): The event decreased firm value.
- The magnitude of \(\Delta W_t\) reflects the economic impact of the marketing event in dollar terms.
By computing \(\Delta W_t\), researchers can translate stock market reactions into tangible financial implications, helping assess the real-world significance of marketing decisions.
Figure 33.3 shows the simulated distribution of dollar wealth change \(\Delta W_t\) across 100 firms drawn from a normal CAR distribution and a uniform pre-event market value.
# Load necessary libraries
library(tidyverse)
# Simulated dataset of event study results
df_event_study <- tibble(
firm_id = 1:100,
# 100 firms
CAR_t = rnorm(100, mean = 0.02, sd = 0.05),
# Simulated CAR values
MKTVAL_0 = runif(100, min = 1e8, max = 5e9) # Market value in dollars
)
# Compute total wealth gain/loss
df_event_study <- df_event_study %>%
mutate(wealth_change = CAR_t * MKTVAL_0)
# Summary statistics of economic impact
summary(df_event_study$wealth_change)
#> Min. 1st Qu. Median Mean 3rd Qu. Max.
#> -217693173 -8830490 42057727 52288170 106928569 545942284
# Histogram of total wealth gain/loss
hist(
df_event_study$wealth_change,
main = "Distribution of Wealth Change from Event",
xlab = "Wealth Change ($)",
col = "blue",
breaks = 30
)
Figure 33.3: Distribution of simulated dollar wealth change from a marketing event across 100 firms, computed as CAR times pre-event market value.
33.6 Testing in Event Studies
33.6.1 Statistical Power in Event Studies
Statistical power refers to the ability to detect a true effect (i.e., identify significant abnormal returns) when one exists.
Power increases with:
- More firms in the sample → reduces variance and increases reliability.
- Fewer days in the event window → avoids contamination from other confounding factors.
Trade-Off:
- A longer event window captures delayed market reactions but risks contamination from unrelated events.
- A shorter event window reduces noise but may miss slow adjustments in stock prices.
Thus, an optimal event window balances precision (avoiding confounds) and completeness (capturing true market reaction).
33.6.2 Parametric Tests
S. J. Brown and Warner (1985) provide evidence that parametric tests perform well even under non-normality, as long as the sample includes at least five securities. This is because the distribution of abnormal returns converges to normality as the sample size increases.
33.6.2.1 Power of Parametric Tests
Kothari and Warner (1997) highlights that the power to detect significant abnormal returns depends on:
- Sample size: More firms improve statistical power.
- Magnitude of abnormal returns: Larger effects are easier to detect.
- Variance of abnormal returns across firms: Lower variance increases power.
33.6.2.2 T-Test for Abnormal Returns
By applying the Central Limit Theorem, we can use the t-test for abnormal returns:
\[ \begin{aligned} t_{CAR} &= \frac{\overline{CAR_{it}}}{\sigma (CAR_{it})/\sqrt{n}} \\ t_{BHAR} &= \frac{\overline{BHAR_{it}}}{\sigma (BHAR_{it})/\sqrt{n}} \end{aligned} \]
Assumptions:
Abnormal returns follow a normal distribution.
Variance is equal across firms.
No cross-sectional correlation in abnormal returns.
If these assumptions do not hold, the t-test will be misspecified, leading to unreliable inference.
Misspecification may occur due to:
Heteroskedasticity (unequal variance across firms).
Cross-sectional dependence (correlation in abnormal returns across firms).
Non-normality of abnormal returns (though event study design often forces normality).
To address these concerns, Patell Standardized Residuals provide a robust alternative.
33.6.2.3 Patell Standardized Residual
Patell (1976) developed the Patell Standardized Residuals (PSR), which standardizes abnormal returns to correct for estimation errors.
Since the market model relies on observations outside the event window, it introduces prediction errors beyond true residuals. PSR corrects for this:
\[ AR_{it} = \frac{\hat{u}_{it}}{s_i \sqrt{C_{it}}} \]
where:
- \(\hat{u}_{it}\) = estimated residual from the market model.
- \(s_i\) = standard deviation of residuals from the estimation period.
- \(C_{it}\) = correction factor accounting for estimation period variation.
The correction factor (\(C_{it}\)) is:
\[ C_{it} = 1 + \frac{1}{T} + \frac{(R_{mt} - \bar{R}_m)^2}{\sum_t (R_{mt} - \bar{R}_m)^2} \]
where:
- \(T\) = number of observations in the estimation period.
- \(R_{mt}\) = market return at time \(t\).
- \(\bar{R}_m\) = mean market return.
This correction ensures abnormal returns are properly scaled, reducing bias from estimation errors.
33.6.3 Non-Parametric Tests
Non-parametric tests do not assume a specific return distribution, making them robust to non-normality and heteroskedasticity.
33.6.3.1 Sign Test
The Sign Test assumes symmetric abnormal returns around zero.
- Null hypothesis (\(H_0\)): Equal probability of positive and negative abnormal returns.
- Alternative hypothesis (\(H_A\)): More positive (or negative) abnormal returns than expected.
# Perform a sign test using binomial test
binom.test(x = sum(CAR > 0), n = length(CAR), p = 0.5)33.6.3.2 Wilcoxon Signed-Rank Test
The Wilcoxon Signed-Rank Test allows for non-symmetry in returns.
Use case: Detects shifts in the distribution of abnormal returns.
More powerful than the sign test when return magnitudes matter.
# Perform Wilcoxon Signed-Rank Test
wilcox.test(CAR, mu = 0)33.6.3.3 Generalized Sign Test
A more advanced sign test, comparing the proportion of positive abnormal returns to historical norms.
33.6.3.4 Corrado Rank Test
The Corrado Rank Test is a rank-based test for abnormal returns.
Advantage: Accounts for cross-sectional dependence.
More robust than the t-test under non-normality.
# Load necessary libraries
library(tidyverse)
# Simulate abnormal returns (CAR)
set.seed(123)
df_returns <- tibble(
firm_id = 1:100, # 100 firms
CAR = rnorm(100, mean = 0.02, sd = 0.05) # Simulated CAR values
)
# Parametric T-Test for CAR
t_test_result <- t.test(df_returns$CAR, mu = 0)
# Non-parametric tests
sign_test_result <- binom.test(sum(df_returns$CAR > 0), n = nrow(df_returns), p = 0.5)
wilcox_test_result <- wilcox.test(df_returns$CAR, mu = 0)
# Print results
list(
T_Test = t_test_result,
Sign_Test = sign_test_result,
Wilcoxon_Test = wilcox_test_result
)
#> $T_Test
#>
#> One Sample t-test
#>
#> data: df_returns$CAR
#> t = 5.3725, df = 99, p-value = 5.159e-07
#> alternative hypothesis: true mean is not equal to 0
#> 95 percent confidence interval:
#> 0.01546417 0.03357642
#> sample estimates:
#> mean of x
#> 0.0245203
#>
#>
#> $Sign_Test
#>
#> Exact binomial test
#>
#> data: sum(df_returns$CAR > 0) and nrow(df_returns)
#> number of successes = 70, number of trials = 100, p-value = 7.85e-05
#> alternative hypothesis: true probability of success is not equal to 0.5
#> 95 percent confidence interval:
#> 0.6001853 0.7875936
#> sample estimates:
#> probability of success
#> 0.7
#>
#>
#> $Wilcoxon_Test
#>
#> Wilcoxon signed rank test with continuity correction
#>
#> data: df_returns$CAR
#> V = 3917, p-value = 1.715e-06
#> alternative hypothesis: true location is not equal to 033.7 Sample in Event Studies
A practical question that often surprises newcomers to event studies is how few observations are typically involved. Marketing and finance applications routinely run on samples that would look anaemic in other empirical traditions, and yet they regularly yield publishable, interpretable results. A glance at the published record gives a sense of the range. Markovitch and Golder (2008) work with 71 events at the small end of the distribution; Wiles, Morgan, and Rego (2012) have a more typical setup with 572 acquisition announcements and 308 disposal announcements; Borah and Tellis (2014) sit at the upper end with 3,552 events. The lesson is not that sample size doesn’t matter, larger samples buy more power and tighter inference, but that the signal in an event study comes from the sharpness of the event window relative to normal-return variation, not from raw \(N\). With clean events and a well-specified normal-return model, a few dozen carefully curated cases can yield results that would survive in a much larger study with noisier identification.
33.8 Confounders in Event Studies
A major challenge in event studies is controlling for confounding events, which could bias the estimation of abnormal returns.
33.8.1 Types of Confounding Events
(McWilliams and Siegel 1997) suggest excluding firms that experience other major events within a two-day window around the focal event. These include:
- Financial announcements: Earnings reports, stock buybacks, dividend changes, IPOs.
- Corporate actions: Mergers, acquisitions, spin-offs, stock splits, debt defaults.
- Executive changes: CEO/CFO resignations or appointments.
- Operational changes: Layoffs, restructurings, lawsuits, joint ventures.
Fornell et al. (2006) recommend:
- One-day event period: The date when Wall Street Journal publishes the ACSI announcement.
- Five-day window (before and after the event) to rule out other news (from PR Newswires, Dow Jones, Business Wires).
Events controlled for include:
M&A, spin-offs, stock splits.
CEO or CFO changes.
Layoffs, restructurings, lawsuits.
A useful data source for identifying confounding events is Capital IQ’s Key Developments, which captures almost all important corporate events.
33.8.2 Should We Exclude Confounded Observations?
A. Sorescu, Warren, and Ertekin (2017) investigated confounding events in short-term event windows using:
RavenPack dataset (2000-2013).
3-day event windows for 3,982 US publicly traded firms.
Key Findings:
- The difference between the full sample and the sample without confounded events was statistically insignificant.
- Conclusion: Excluding confounded observations may not be necessary in short-term event studies.
Why?
- Selection bias risk: Researchers may selectively exclude events, introducing bias.
- Increasing exclusions over time: As time progresses, more events need to be excluded, reducing statistical power.
- Short-term windows minimize confounder effects.
33.9 Biases in Event Studies
Event studies are subject to several biases that can affect the estimation of abnormal returns, the validity of test statistics, and the interpretation of results. The biases below are not exhaustive, and additional concerns specific to particular event types or markets may arise; the discussion focuses on the ones that recur often enough to warrant routine attention.
33.9.1 Timing Bias: Different Market Closing Times
When firms in the sample trade on exchanges in different time zones, the very notion of “the day of the event” becomes ambiguous. Campbell et al. (1998) flag this issue: a closing price recorded at 4 p.m. New York time and another at 4 p.m. Tokyo time correspond to substantially different information sets, and aggregating them as if they were contemporaneous obscures the true price reaction. The bias is especially acute for firms cross-listed on multiple exchanges or for events that release news during a specific time zone’s trading hours, after some markets have closed and before others open.
The standard fixes line up directly with the source of the problem. Use synchronized closing prices wherever possible, drawing on a single reference exchange or on intraday quotes recorded at a uniform timestamp. When that is not feasible, define the event window relative to the firm’s primary trading exchange and accept that the resulting estimate is conditional on that choice. In multi-exchange settings, reporting results separately by primary listing is a useful robustness exercise: substantial divergence is a flag that timing alignment is doing real work.
33.9.2 Upward Bias in Cumulative Abnormal Returns
A subtler bias arises in the aggregation of daily abnormal returns into cumulative returns. The mechanism is microstructural rather than statistical: transaction prices recorded at the bid or the ask, rather than at the midpoint, introduce small jumps that the abnormal-return calculation reads as price movement. Liquidity constraints amplify the effect, because thinly traded stocks tend to bounce between bid and ask prices in patterns that look like genuine return innovations. When these microstructural jumps are aggregated over a multi-day event window, the cumulative abnormal return inherits a small but systematic upward bias.
The fixes operate either at the input or at the inference stage. At the input stage, replace raw transaction prices with volume-weighted average prices (VWAP) or with bid-ask midpoints, both of which average out the microstructural noise. At the inference stage, apply heteroskedasticity-robust or microstructure-corrected standard errors, which keep the point estimate unchanged but widen the confidence interval to reflect the additional variance introduced by the bias. For events with very short windows (single trading day), the input-stage fix is more important; for longer windows where the noise averages out anyway, the inference-stage fix usually suffices.
33.9.3 Cross-Sectional Dependence Bias
Cross-sectional dependence in returns biases standard-deviation estimates downward, which in turn inflates test statistics whenever multiple firms experience the event on the same date. MacKinlay (1997) flagged the issue early on, noting that the problem becomes acute when firms in the same industry or market share event dates and so face common shocks that the independence assumption simply cannot accommodate. Wiles, Morgan, and Rego (2012) document the consequences empirically: in concentrated industries, the dependence is severe enough to materially inflate test statistics, and the apparent significance of an event can dissolve once the correction is applied.
Two corrections are standard in the literature, and they target the bias from different angles. The Calendar-Time Portfolio Abnormal Returns (CTAR) approach (Jaffe 1974) reorganizes the data into calendar-time portfolios so that all firms experiencing the event on the same day enter as a single portfolio observation rather than as multiple correlated observations. The dependence problem disappears by construction because there is no longer cross-firm variation within a portfolio. The time-series standard-error correction of S. J. Brown and Warner (1980) takes the opposite route: keep the firm-level structure but estimate the variance from the time-series of the cross-sectional aggregate, capturing the dependence in the variance estimate rather than averaging it away. Both approaches are well-tested in the literature; the choice between them often comes down to whether the analyst wants firm-level attribution (Brown-Warner) or pooled portfolio-level inference (Jaffe).
# Load required libraries
library(sandwich) # For robust standard errors
library(lmtest) # For hypothesis testing
# Simulated dataset
set.seed(123)
df_returns <- data.frame(
event_id = rep(1:100, each = 10),
firm_id = rep(1:10, times = 100),
abnormal_return = rnorm(1000, mean = 0.02, sd = 0.05)
)
# Cross-sectional dependence adjustment using clustered standard errors
model <- lm(abnormal_return ~ 1, data = df_returns)
coeftest(model, vcov = vcovCL(model, cluster = ~event_id))
#>
#> t test of coefficients:
#>
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) 0.0208064 0.0014914 13.951 < 2.2e-16 ***
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 133.9.4 Sample Selection Bias
Event studies often suffer from self-selection bias because firms choose to undertake the events they undertake (issuing equity, announcing an acquisition, recalling a product) on the basis of private information that the analyst does not observe. The firm’s choice is, in effect, a treatment-assignment mechanism that depends on unobservables, and the resulting comparison between event firms and non-event firms reads partly as a treatment effect and partly as the difference in the unobservables that drove the decision. This is the canonical omitted-variable bias problem, with the omitted variable being whatever private information the firm acted on.
33.9.5 Corrections for Sample Selection Bias
Several corrections are available, and the right one depends on the institutional details of the event and on what auxiliary information is available.
The Heckman two-stage model (Acharya 1993) is the parametric approach. A first-stage probit predicts the probability of experiencing the event from observable firm characteristics, and the resulting inverse-Mills ratio enters the second-stage regression of abnormal returns as a control for the unobserved selection mechanism. The strength of the approach is that it has a clear identification logic and a familiar implementation; the weakness is that it requires an exclusion restriction, a variable that affects the probability of the event but not the abnormal return directly, and such instruments are notoriously hard to find in event-study settings.
Counterfactual-observation methods are the non-parametric alternative. Two are common in event-study work. Propensity-score matching pairs each event firm with a non-event firm that has similar observable characteristics, on the logic that conditional on observables the event is as good as random. Switching regression explicitly models the two regimes (event vs. no event) jointly with the selection mechanism, allowing for unobserved heterogeneity in the relationship between firm characteristics and outcomes across the two regimes. Both are useful when the available observables are rich enough to capture the bulk of the selection process.
The remainder of this section walks through each correction in turn, with a code example for the Heckman and propensity-score-matching cases.
- Heckman Selection Model
A Heckman selection model can be used when private information influences both event participation and abnormal returns.
Examples: Y. Chen, Ganesan, and Liu (2009); Wiles, Morgan, and Rego (2012); Fang, Lee, and Yang (2015)
Steps:
First Stage (Selection Equation): Model the firm’s probability of experiencing the event using a Probit regression.
Second Stage (Outcome Equation): Model abnormal returns, controlling for the estimated Mills ratio (\(\lambda\)).
# Load required libraries
library(sampleSelection)
# Simulated dataset for Heckman model
set.seed(123)
df_heckman <- data.frame(
firm_id = 1:500,
event = rbinom(500, 1, 0.3), # Event occurrence (selection)
firm_size = runif(500, 1, 10), # Firm characteristic
abnormal_return = rnorm(500, mean = 0.02, sd = 0.05)
)
# Introduce selection bias by correlating firm_size with event occurrence
df_heckman$event[df_heckman$firm_size > 7] <- 1
# Heckman Selection Model
heckman_model <- selection(
selection = event ~ firm_size, # Selection equation
outcome = abnormal_return ~ firm_size, # Outcome equation
data = df_heckman
)
# Summary of Heckman model
summary(heckman_model)
#> --------------------------------------------
#> Tobit 2 model (sample selection model)
#> Maximum Likelihood estimation
#> Newton-Raphson maximisation, 6 iterations
#> Return code 8: successive function values within relative tolerance limit (reltol)
#> Log-Likelihood: 165.4579
#> 500 observations (239 censored and 261 observed)
#> 6 free parameters (df = 494)
#> Probit selection equation:
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) -1.75936 0.15793 -11.14 <2e-16 ***
#> firm_size 0.33933 0.02776 12.22 <2e-16 ***
#> Outcome equation:
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) 0.006025 0.040359 0.149 0.881
#> firm_size 0.001311 0.004205 0.312 0.755
#> Error terms:
#> Estimate Std. Error t value Pr(>|t|)
#> sigma 0.049048 0.002836 17.297 <2e-16 ***
#> rho 0.188195 0.421944 0.446 0.656
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#> --------------------------------------------Interpretation
If the Mills ratio (\(\lambda\)) is significant, it indicates that private information affects CARs.
Weak instruments can lead to multicollinearity, making the second-stage estimates unreliable.
- Propensity Score Matching
PSM matches event firms with similar non-event firms, controlling for selection bias.
Examples of PSM in Finance and Marketing:
# Load required libraries
library(MatchIt)
# Simulated dataset
set.seed(123)
df_psm <- data.frame(
firm_id = 1:1000,
event = rbinom(1000, 1, 0.5), # 50% of firms experience an event
firm_size = runif(1000, 1, 10),
market_cap = runif(1000, 100, 10000)
)
# Propensity score matching (PSM)
match_model <- matchit(event ~ firm_size + market_cap, data = df_psm, method = "nearest")
# Summary of matched sample
summary(match_model)
#>
#> Call:
#> matchit(formula = event ~ firm_size + market_cap, data = df_psm,
#> method = "nearest")
#>
#> Summary of Balance for All Data:
#> Means Treated Means Control Std. Mean Diff. Var. Ratio eCDF Mean
#> distance 0.4987 0.4875 0.2093 1.0656 0.0602
#> firm_size 5.2627 5.6998 -0.1683 1.0530 0.0494
#> market_cap 5208.5283 4868.5828 0.1163 1.0483 0.0359
#> eCDF Max
#> distance 0.1152
#> firm_size 0.0902
#> market_cap 0.0713
#>
#> Summary of Balance for Matched Data:
#> Means Treated Means Control Std. Mean Diff. Var. Ratio eCDF Mean
#> distance 0.4987 0.4898 0.1668 1.1170 0.0489
#> firm_size 5.2627 5.6182 -0.1369 1.0693 0.0404
#> market_cap 5208.5283 4949.8521 0.0885 1.0594 0.0283
#> eCDF Max Std. Pair Dist.
#> distance 0.1034 0.1673
#> firm_size 0.0872 0.6549
#> market_cap 0.0649 0.9168
#>
#> Sample Sizes:
#> Control Treated
#> All 507 493
#> Matched 493 493
#> Unmatched 14 0
#> Discarded 0 0
# Extract matched data
matched_data <- match.data(match_model)Advantages of PSM
Controls for observable differences between event and non-event firms.
Reduces selection bias while maintaining a valid control group.
- Switching Regression
A Switching Regression Model accounts for selection on unobservables using instrumental variables.
- Example: Cao and Sorescu (2013) applied switching regression to compare two outcomes while correcting for selection bias.
33.10 Long-run Event Studies
Long-horizon event studies analyze the long-term impact of corporate events on stock prices. These studies commonly assume that the distribution of abnormal returns has a mean of zero (A. Sorescu, Warren, and Ertekin 2017, 192). Moreover, A. Sorescu, Warren, and Ertekin (2017) provide evidence that samples with and without confounding events yield similar results.
However, long-run event studies face several methodological challenges:
- Systematic biases over time: Estimation errors can accumulate over long periods.
- Sensitivity to model specification: The choice of asset pricing models can influence results.
Long-run event studies typically use event windows of 12 to 60 months (Loughran and Ritter 1995; Brav and Gompers 1997).
There are three primary methods for measuring long-term abnormal stock returns:
- Buy-and-Hold Abnormal Returns (BHAR)
- Long-term Cumulative Abnormal Returns (LCARs)
- Calendar-time Portfolio Abnormal Returns (CTARs), also known as Jensen’s Alpha, which better handles cross-sectional dependence and is less sensitive to asset pricing model misspecification.
Types of Events Analyzed in Long-run Studies
-
Unexpected changes in firm-specific variables
These events are typically not announced, may not be immediately visible to all investors, and their impact on firm value is complex. Examples include:- The effect of customer satisfaction scores on firm value (Jacobson and Mizik 2009).
- Unexpected changes in marketing expenditures and their potential mispricing effects (M. Kim and McAlister 2011).
-
Events with complex consequences
Investors may take time to fully incorporate the information into stock prices. For example:- The long-term impact of acquisitions depends on post-merger integration (A. B. Sorescu, Chandy, and Prabhu 2007).
Below is an example using the crseEventStudy package, which calculates standardized abnormal returns:
library(crseEventStudy)
# Example using demo data from the package
data(demo_returns)
SAR <- sar(event = demo_returns$EON,
control = demo_returns$RWE,
logret = FALSE)
mean(SAR)
#> [1] 0.00687019633.10.1 Buy-and-Hold Abnormal Returns (BHAR)
BHAR is one of the most widely used methods in long-term event studies. It involves constructing a portfolio of benchmark stocks that closely match event firms over the same period and then comparing their returns. The approach was popularised by Loughran and Ritter (1995) in their work on long-run post-issue performance, refined methodologically by Barber and Lyon (1997), and given a careful treatment of the inferential problems that long horizons introduce by Lyon, Barber, and Tsai (1999), these three papers are the entry points to the literature, and the procedural choices below largely reflect their accumulated guidance.
BHAR measures returns from:
Buying stocks in event firms.
Shorting stocks in similar non-event firms.
Since cross-sectional correlations can inflate t-statistics, BHAR’s rank order remains reliable even if absolute significance levels are affected (Markovitch and Golder 2008; A. B. Sorescu, Chandy, and Prabhu 2007).
To construct the benchmark portfolio, firms are matched based on:
Size
Book-to-market ratio
Momentum
Matching strategies vary across studies. Below are two common procedures:
- (Barber and Lyon 1997) approach
Each July, all common stocks in the CRSP database are classified into ten deciles based on market capitalization from the previous June.
Within each size decile, firms are further grouped into five quintiles based on their book-to-market ratios as of the prior December.
The benchmark portfolio consists of non-event firms that fit these criteria.
- (Wiles et al. 2010) approach
Firms in the same two-digit SIC code with market values between 50% and 150% of the focal firm are selected.
From this subset, the 10 firms with the closest book-to-market ratios form the benchmark portfolio.
With the abnormal-return primitive \(AR_{it}\) and the arithmetic accumulator \(CAR_{it}\) already defined in Section 33 (Step 4), the long-run object of interest swaps the sum for a product over the post-event window:
\[ BHAR_{t=1}^{T} = \prod_{t=1}^{T} (1 + R_{it}) - \prod_{t=1}^{T} (1 + E(R_{it})) \]
Unlike CAR, which is arithmetic, BHAR is geometric.
In short-term studies, differences between CAR and BHAR are minimal.
In long-term studies, the discrepancy is significant. For instance, Barber and Lyon (1997) show that when annual BHAR exceeds 28%, it dramatically surpasses CAR.
To avoid favoring recent events, researchers in cross-sectional event studies typically treat all events equally when assessing their impact on the stock market over time. This approach helps in identifying abnormal changes in stock prices, particularly when analyzing a series of unplanned events.
However, long-run event studies face several biases that can distort abnormal return calculations:
- Construct Benchmark Portfolios with Fixed Constituents
One recommended approach is to form benchmark portfolios that do not change their constituent firms over time (Mitchell and Stafford 2000). This helps mitigate the following biases:
New Listing Bias
Newly public companies often underperform relative to a balanced market index (Ritter 1991). Including these firms in event studies may distort long-term return expectations. This issue, termed new listing bias, was first identified by (Barber and Lyon 1997).Rebalancing Bias
Regularly rebalancing an equal-weighted portfolio can lead to overestimated long-term returns. This is because the process systematically sells winning stocks and buys underperformers, which tends to skew buy-and-hold abnormal returns downward (Barber and Lyon 1997).Value-Weight Bias
Value-weighted portfolios, which assign higher weights to larger market capitalization stocks, may overestimate BHARs. This approach mimics an active strategy that continuously buys winners and sells underperformers, which inflates long-run return estimates.
- Buy-and-Hold Without Annual Rebalancing
Another method involves holding an initial portfolio fixed throughout the investment period. In this approach, returns are compounded, and the average is calculated across all securities:
\[ \Pi_{t = s}^{T} (1 + E(R_{it})) = \sum_{i=s}^{n_t} \left( w_{is} \prod_{t=1}^{T} (1 + R_{it}) \right) \]
where:
\(T\) = total investment period,
\(R_{it}\) = return on security \(i\) at time \(t\),
\(n_t\) = number of securities in the portfolio,
\(w_{is}\) = initial weight of firm \(i\) in the portfolio at period \(s\) (either equal-weighted or value-weighted).
Key Characteristics of This Approach
-
No Monthly Adjustments
The portfolio remains fixed based on stocks available at time \(s\), meaning:- No new stocks are added after period \(s\).
- No rebalancing occurs each period.
Avoids Rebalancing Bias
Since there is no forced buying or selling, distortions due to rebalancing are minimized.Market-Weight Adjustment is Required
Since value-weighted portfolios favor larger firms, adjustments may be necessary to prevent recently listed firms from exerting excessive influence on portfolio returns.
- The choice between equal-weighted and value-weighted portfolios affects results:
- Equal-weighted portfolios ensure each firm contributes equally.
- Value-weighted portfolios reflect real-world investment scenarios but may be skewed toward larger firms.
- Researchers should define minimum inclusion criteria (e.g., stocks must trade for at least 12 months post-event) to filter out firms with insufficient return data.
For empirical research, Wharton Research Data Services (WRDS) provides an automated tool for computing Buy-and-Hold Abnormal Returns. This tool allows researchers to generate all types of BHAR measures based on different weighting and rebalancing approaches:
- Equal-weighted vs. Value-weighted portfolios
- With vs. Without annual rebalancing
The WRDS platform enables users to upload their own event data and apply these methodologies efficiently. More details can be found at WRDS Long-Run Event Study.
The WRDS tool provides several options for customizing event study settings, summarised in Table 33.8.
| Parameter | Description |
|---|---|
| MINWIN | The minimum number of months a firm must trade after the event to be included in the study. |
| MAXWIN | The maximum number of months considered in the event study. |
| MONTH | The event window length (e.g., 12, 24, or 36 months) for BHAR calculation. |
If a firm’s monthly returns are missing during the selected event window, matching portfolio returns are used to fill in the gaps. This ensures that BHAR calculations remain consistent even when individual firm data is incomplete.
33.10.2 Long-term Cumulative Abnormal Returns (LCARs)
Long-term Cumulative Abnormal Returns (LCARs) measure the total abnormal return of an event firm over an extended period post-event. Unlike Buy-and-Hold Abnormal Returns, which use compounding, LCARs sum up abnormal returns over time.
This method is widely used in long-run event studies and is particularly useful for examining how an event’s impact evolves gradually rather than instantaneously.
The LCAR for firm \(i\) over the post-event horizon \((1,T)\) is given by (A. B. Sorescu, Chandy, and Prabhu 2007):
\[ LCAR_{iT} = \sum_{t = 1}^{T} (R_{it} - R_{pt}) \]
where:
\(R_{it}\) = Rate of return of stock \(i\) in month \(t\).
\(R_{pt}\) = Rate of return on the counterfactual (benchmark) portfolio in month \(t\).
LCARs aggregate monthly abnormal returns to capture the cumulative effect of an event over time.
33.10.2.1 Key Considerations in Using LCARs
- Benchmark Portfolio Selection
The choice of counterfactual portfolio \(R_{pt}\) is critical, as it serves as a reference point for detecting abnormal performance. Common benchmarks include:
Size and book-to-market matched portfolios
Firms are grouped based on market capitalization and book-to-market ratio to control for firm characteristics.Industry-matched portfolios
Firms within the same industry (e.g., 2-digit SIC code) provide a relevant comparison.Market model expectations
Expected returns are estimated using asset pricing models such as the CAPM or Fama-French 3-factor model.
- Event Window Length
Long-term event studies use windows ranging from 12 to 60 months (Loughran and Ritter 1995; Brav and Gompers 1997). A longer window captures the full market reaction but increases the risk of contamination from unrelated events.
- Statistical Significance Issues
Since LCARs use a simple summation of abnormal returns, they can suffer from:
- Cross-sectional dependence: Abnormal returns across firms may be correlated, inflating t-statistics.
- Variance drift: The standard deviation of cumulative returns grows over time, complicating inference.
To correct these biases, researchers often use:
- Bootstrapping methods
- Calendar-time portfolio approaches (e.g., Jensen’s Alpha)
- Skewness-adjusted t-tests (Lyon, Barber, and Tsai 1999)
Table 33.9 contrasts the two long-run abnormal return aggregations.
| Feature | LCAR | BHAR |
|---|---|---|
| Computation | Sum of abnormal returns | Product of abnormal returns |
| Return Aggregation | Arithmetic | Geometric |
| Main Issue | Variance drift | Rebalancing bias |
| Best for | Identifying gradual changes in stock performance | Capturing compounding effects |
In short-term studies, LCAR and BHAR tend to yield similar results, but in long-term studies, BHAR amplifies the impact of extreme returns, whereas LCAR provides a more linear view.
Figure 33.5 plots simulated long-run cumulative abnormal return paths for 50 event firms over a 60-month event window.
# Load necessary packages
library(tidyverse)
library(ggplot2)
# Simulate stock returns and benchmark portfolio returns
set.seed(123)
months <- 60 # 5-year event window
firms <- 50 # Number of event firms
# Generate random stock returns (normally distributed)
stock_returns <-
matrix(rnorm(months * firms, mean = 0.01, sd = 0.05),
nrow = months,
ncol = firms)
# Generate benchmark portfolio returns
benchmark_returns <- rnorm(months, mean = 0.009, sd = 0.03)
# Compute LCAR for each firm
LCARs <-
apply(stock_returns, 2, function(stock)
cumsum(stock - benchmark_returns))
# Convert to data frame for visualization
LCAR_df <- as.data.frame(LCARs) %>%
mutate(Month = 1:months) %>%
pivot_longer(-Month, names_to = "Firm", values_to = "LCAR")
# Plot LCAR trajectories
ggplot(LCAR_df, aes(x = Month, y = LCAR, group = Firm)) +
geom_line(alpha = 0.3) +
theme_minimal() +
labs(
title = "Long-term Cumulative Abnormal Returns (LCARs)",
x = "Months Since Event",
y = "Cumulative Abnormal Return",
caption = "Each line represents an event firm's LCAR trajectory."
)
Figure 33.5: Simulated long-run cumulative abnormal return (LCAR) trajectories for 50 firms over a 60-month event window, with each line representing one firm’s cumulative abnormal return relative to a benchmark portfolio.
33.10.3 Calendar-time Portfolio Abnormal Returns (CTARs)
The Calendar-time Portfolio Abnormal Returns (CTARs) method, also known as Jensen’s Alpha approach, is widely used in long-run event studies to address cross-sectional dependence among firms experiencing similar events. Unlike BHAR or LCAR, which focus on individual stock returns, CTARs evaluate portfolio-level abnormal returns over time.
This method follows the strict procedure outlined in Wiles et al. (2010) and has key advantages:
Controls for cross-sectional correlation by aggregating event firms into portfolios.
Reduces model misspecification biases by relying on time-series regressions instead of individual firm-level return calculations.
33.10.3.1 Constructing the Calendar-time Portfolio
- Portfolio Formation
- A portfolio is constructed for every day in the calendar time (including all firms that experience an event on that day).
- Securities in each portfolio are equally weighted to avoid bias from firm size differences.
- A portfolio is constructed for every day in the calendar time (including all firms that experience an event on that day).
- Compute the Average Abnormal Return for Each Portfolio
For a given portfolio \(P\) on day \(t\):
\[ AAR_{Pt} = \frac{\sum_{i=1}^S AR_i}{S} \]
where:
\(S\) = Number of stocks in portfolio \(P\).
\(AR_i\) = Abnormal return for stock \(i\) in the portfolio.
-
Calculate the Standard Deviation of AAR over the Preceding \(k\) Days
The time-series standard deviation of \(AAR_{Pt}\), denoted as \(SD(AAR_{Pt})\), is calculated using the preceding \(k\) days (rolling window), assuming independence over time.
Standardize the Average Abnormal Return
\[ SAAR_{Pt} = \frac{AAR_{Pt}}{SD(AAR_{Pt})} \]
- Compute the Average Standardized Abnormal Return (ASAAR)
The standardized residuals across all portfolios are averaged across the full calendar time:
\[ ASAAR = \frac{1}{n} \sum_{t=1}^{255} SAAR_{Pt} \times D_t \]
where:
\(D_t = 1\) when at least one security is in portfolio \(P_t\), otherwise \(D_t = 0\).
\(n\) is the number of days where at least one firm is in the portfolio, defined as:
\[ n = \sum_{t=1}^{255} D_t \]
- Compute the Cumulative Average Standardized Abnormal Return (CASSAR)
The cumulative impact of events over a time horizon \(S_1\) to \(S_2\) is given by:
\[ CASSAR_{S_1, S_2} = \sum_{t=S_1}^{S_2} ASAAR \]
- Compute the Test Statistic
If ASAAR values are independent over time, the standard deviation of the cumulative metric is:
\[ \sqrt{S_2 - S_1 + 1} \]
Thus, the test statistic for assessing statistical significance is:
\[ t = \frac{CASSAR_{S_1, S_2}}{\sqrt{S_2 - S_1 + 1}} \]
33.10.3.2 Limitations of the CTAR Method
While CTAR offers robust cross-sectional controls, it has notable limitations:
- Cannot Examine Individual Stock Differences
- CTAR only evaluates portfolio-level differences, masking firm-level variations.
- A workaround is to construct multiple portfolios based on relevant firm characteristics (e.g., size, book-to-market, industry) and compare their intercepts.
- Low Statistical Power
- CTAR has been criticized for low power (i.e., high Type II error rates) (Loughran and Ritter 2000).
- Detecting significant abnormal returns requires a large number of event firms and a sufficiently long time-series.
33.11 Aggregation
33.11.1 Over Time
To assess the impact of events on stock performance over time, we calculate the Cumulative Abnormal Return (CAR) for the event windows.
Hypotheses:
- \(H_0\): The standardized cumulative abnormal return (SCAR) for stock \(i\) is 0 (i.e., the event has no effect on stock performance).
- \(H_1\): The SCAR is not 0 (i.e., the event does have an effect on stock performance).
33.11.2 Across Firms and Over Time
In addition to evaluating CAR for individual stocks, we may want to aggregate results across multiple firms to determine whether events systematically affect stock prices.
Additional Assumptions:
- Uncorrelated Abnormal Returns: The abnormal returns of different stocks are assumed to be uncorrelated. This is a strong assumption, but it holds reasonably well if event windows for different stocks do not overlap.
- Overlapping Event Windows: If event windows do overlap, follow the methodology proposed by Bernard (1987) and Schipper and Thompson (1983), Schipper and Smith (1983).
Hypotheses:
- \(H_0\): The mean abnormal return across all firms is 0 (i.e., there is no systematic effect of the event).
- \(H_1\): The mean abnormal return across all firms is different from 0 (i.e., the event has a systematic effect).
33.11.3 Statistical Tests
The parametric (t-test, Patell Standardized Residual) and non-parametric (sign, Wilcoxon, generalized sign, Corrado rank) tests catalogued in Section 33.6 carry over directly to the cross-firm aggregate, with the choice between aggregating CAR or SCAR turning on whether the true abnormal-return variance scales with firm-level return variance (favoring CAR) or is roughly constant across firms (favoring SCAR).
33.12 Heterogeneity in the Event Effect
The impact of an event on stock performance can vary significantly across firms due to firm-specific or event-specific characteristics. We model this heterogeneity using the following regression framework:
\[ y = X \theta + \eta \]
where:
- \(y\) = Cumulative Abnormal Return (CAR) for a given event window.
- \(X\) = Matrix of firm- or event-specific characteristics that explain heterogeneity in the event effect.
- \(\theta\) = Vector of coefficients capturing the impact of these characteristics on abnormal returns.
- \(\eta\) = Error term, capturing unobserved factors.
Selection bias can arise if firm characteristics influence both the likelihood of experiencing the event and the magnitude of abnormal returns. One common issue is investor anticipation:
Example: Larger firms might benefit more from an event, leading investors to preemptively price in the expected effect, potentially distorting CAR measurements.
This can result in an endogeneity problem, where expected returns are systematically related to firm characteristics.
To correct for this issue, White’s heteroskedasticity-consistent \(t\)-statistics should be used. This provides lower bounds for the true significance of coefficient estimates by accounting for heteroskedasticity in the regression residuals.
Key Point: Even if the average CAR is not significantly different from zero, analyzing heterogeneity remains essential, particularly when CAR variance is high (Boyd, Chandy, and Cunha Jr 2010).
33.12.1 Common Variables Affecting CAR in Marketing and Finance
Event effects on stock returns can be influenced by various firm-specific and market-specific factors. The following variables are commonly examined in event studies, as summarized in A. Sorescu, Warren, and Ertekin (2017) (Table 4).
Firm-Specific Characteristics
-
Firm Size
- Finance Literature: Typically negatively correlated with abnormal returns.
- Marketing Literature: Results are mixed, suggesting different dynamics.
- Interpretation: Large firms may have less information asymmetry, leading to smaller stock reactions.
-
Number of Event Occurrences
- A firm that frequently experiences similar events may see diminishing stock market reactions over time.
-
R&D Expenditure
- Higher R&D investment often signals long-term innovation potential but may also increase risk, affecting abnormal returns.
-
Advertising Expense
- Can enhance brand equity and consumer perception, leading to a stronger stock price response to events.
-
Marketing Investment (SG&A - Selling, General & Administrative Expenses)
- A proxy for strategic spending on market development.
- High marketing investment may drive higher abnormal returns if perceived as value-enhancing.
-
Financial Leverage (Debt-to-Equity Ratio)
- High leverage can amplify risk, leading to more pronounced market reactions to events.
-
Book-to-Market Ratio
- A fundamental indicator of valuation.
- High book-to-market firms (value stocks) may respond differently to events compared to low book-to-market firms (growth stocks).
-
Return on Assets (ROA)
- A measure of firm profitability.
- Higher ROA firms may be less susceptible to negative shocks.
-
Free Cash Flow
- High free cash flow can signal financial flexibility, potentially mitigating negative event impacts.
-
Sales Growth
- A proxy for firm momentum.
- Higher growth firms may exhibit stronger abnormal returns following positive events.
-
Firm Age
- Younger firms may experience higher abnormal returns due to greater investor uncertainty and information asymmetry.
Industry-Specific & Market-Level Characteristics
-
Industry Concentration (Herfindahl-Hirschman Index - HHI, Number of Competitors)
- High industry concentration (fewer competitors) can reduce competitive pressure, leading to stronger abnormal returns.
-
Market Share
- Firms with higher market share may experience weaker abnormal returns due to already-established dominance.
-
Market Size (Total Sales Volume within the Firm’s SIC Code)
- A measure of industry attractiveness.
- Events occurring in larger markets may have muted effects due to broader investor diversification.
-
Marketing Capability
- Firms with stronger marketing capabilities may better leverage events for long-term brand and revenue growth, influencing CAR.
33.13 Expected Return Calculation
Expected return models are essential for estimating abnormal returns in event studies. These models help separate normal stock price movements from those caused by specific events.
33.13.1 Statistical Models for Expected Returns
Statistical models rely on assumptions about the behavior of returns, often assuming stable distributions (Owen and Rabinovitch 1983). These models do not impose economic constraints but instead focus on statistical properties of returns.
33.13.1.1 Constant Mean Return Model
The simplest statistical model assumes that a stock’s expected return is simply its historical mean return:
\[ Ra_{it} = R_{it} - \bar{R}_i \]
where:
- \(R_{it}\) = observed return of stock \(i\) in period \(t\)
- \(\bar{R}_i\) = mean return of stock \(i\) over the estimation period
- \(Ra_{it}\) = abnormal return in period \(t\) (i.e., deviation from historical average)
Assumptions:
- Returns revert to their mean over time (i.e., they follow a stable mean-reverting process).
- This assumption is questionable, as market conditions evolve dynamically.
Empirical Note:
The constant mean return model typically delivers similar results to more complex models since the variance of abnormal returns is not substantially reduced when using more sophisticated statistical approaches (S. J. Brown and Warner 1985).
33.13.1.2 Market Model
A widely used alternative to the constant mean model is the market model, which assumes that stock returns are linearly related to market returns:
\[ R_{it} = \alpha_i + \beta_i R_{mt} + \epsilon_{it} \]
where:
- \(R_{it}\) = return of stock \(i\) in period \(t\)
- \(R_{mt}\) = market return in period \(t\) (e.g., S&P 500 index)
- \(\alpha_i\) = stock-specific intercept (capturing average return not explained by the market)
- \(\beta_i\) = systematic risk (market beta) of stock \(i\)
- \(\epsilon_{it}\) = zero-mean error term with variance \(\sigma^2\), capturing idiosyncratic risk
Notes on Implementation:
- The market return (\(R_{mt}\)) is typically proxied using:
- S&P 500 index
- CRSP value-weighted index
- CRSP equal-weighted index
- If \(\beta_i = 0\), the market model reduces to the constant mean return model.
Key Insight:
The better the fit of the market model, the lower the variance of abnormal returns, making it easier to detect event effects.
Robust Estimation:
- To account for heteroskedasticity and autocorrelation, it is recommended to use the Generalized Method of Moments (GMM) for estimation.
33.13.1.3 Fama-French Multifactor Models
The Fama-French family of models extends the market model by incorporating additional factors that capture systematic risks beyond market exposure.
Key Considerations:
- There is a distinction between using total return and excess return as the dependent variable.
- The correct specification involves excess returns for both individual stocks and the market portfolio (Fama and French 2010, 1917).
Interpretation of \(\alpha_i\):
\(\alpha_i\) represents the abnormal return, i.e., the return that is unexplained by the model.
33.13.1.3.1 Fama-French Three-Factor Model (FF3)
\[ \begin{aligned} E(R_{it}|X_t) - r_{ft} &= \alpha_i + \beta_{1i} (E(R_{mt}|X_t )- r_{ft}) \\ &+ b_{2i} SML_t + b_{3i} HML_t \end{aligned} \]
where:
- \(r_{ft}\) = risk-free rate (e.g., 3-month Treasury bill)
- \(R_{mt}\) = market return (e.g., S&P 500)
- \(SML_t\) = size factor (returns on small-cap stocks minus large-cap stocks)
- \(HML_t\) = value factor (returns on high book-to-market stocks minus low book-to-market stocks)
33.13.1.3.2 Fama-French Four-Factor Model (FF4)
(Carhart 1997) extends FF3 by adding a momentum factor:
\[ \begin{aligned} E(R_{it}|X_t) - r_{ft} &= \alpha_i + \beta_{1i} (E(R_{mt}|X_t )- r_{ft}) \\ &+ b_{2i} SML_t + b_{3i} HML_t + b_{4i} UMD_t \end{aligned} \]
where:
- \(UMD_t\) = momentum factor (returns of high past-return stocks minus low past-return stocks)
Practical Application in Marketing:
(A. Sorescu, Warren, and Ertekin 2017, 195) recommends:
Market Model for short-term event windows.
Fama-French Model for long-term windows.
However, the statistical properties of the FF model for daily event studies remain untested.
33.13.2 Economic Models for Expected Returns
Economic models impose theoretical constraints on expected returns based on equilibrium asset pricing theory. The two most widely used models are:
33.13.2.1 Capital Asset Pricing Model (CAPM)
CAPM is derived from modern portfolio theory and assumes that expected returns are determined solely by market risk:
\[ E(R_i) = R_f + \beta_i (E(R_m) - R_f) \]
where:
- \(E(R_i)\) = expected return of stock \(i\)
- \(R_f\) = risk-free rate
- \(E(R_m) - R_f\) = market risk premium (excess return of market portfolio)
- \(\beta_i\) = firm-specific market beta (systematic risk measure)
Key Assumption:
Investors hold the market portfolio, and only systematic risk (beta) matters.
33.13.2.2 Arbitrage Pricing Theory (APT)
APT generalizes CAPM by allowing multiple risk factors to drive expected returns:
\[ R = R_f + \Lambda f + \epsilon \]
where:
- \(\Lambda\) = factor loadings (sensitivities to risk factors)
-
\(f \sim N(\mu, \Omega)\) = vector of risk factors
- \(\mu\) = expected risk premiums
- \(\Omega\) = factor covariance matrix
- \(\epsilon \sim N(0, \Psi)\) = idiosyncratic error term
APT vs. CAPM:
CAPM assumes a single factor (market risk).
APT allows multiple systematic factors, making it more flexible for empirical applications.
Summary: Model Comparison
Table 33.10 compares the expected-return models commonly used to compute abnormal returns.
| Model | Key Assumptions | Factors Considered | Best Use Case |
|---|---|---|---|
| Constant Mean Return | Mean-reverting returns | None | Simple event studies |
| Market Model | Linear relationship with market returns | Market factor (\(R_m\)) | Short-term studies |
| Fama-French (FF3) | Size & value factors matter | Market, Size (SML), Value (HML) | Medium- to long-term studies |
| Fama-French (FF4) | Momentum also matters | Market, Size, Value, Momentum | Momentum-driven strategies |
| CAPM | Market is the only risk factor | Market factor (\(R_m\)) | Classic asset pricing |
| APT | Multiple systematic risks matter | Market + Other Factors | Flexible risk modeling |
33.14 Application of Event Study
Several R packages facilitate event studies; Table 33.11 lists the most widely used.
| Package | Description |
|---|---|
eventstudies |
Computes abnormal returns and visualizes event impact |
erer |
Implements event study methodology in economics and finance |
EventStudy |
Provides API-based event study tools (requires subscription) |
AbnormalReturns |
Implements market model and FF3 abnormal return calculations |
PerformanceAnalytics |
Provides tools for risk and performance measurement |
estudy2 |
A flexible event study implementation |
To install these packages, run:
install.packages(
c(
"eventstudies",
"erer",
"EventStudy",
"AbnormalReturns",
"PerformanceAnalytics",
"tidyquant",
"tidyverse"
)
)33.14.1 Sorting Portfolios for Expected Returns
A common approach in finance is to sort stocks into portfolios based on firm characteristics such as size and book-to-market (B/M) ratio. This method helps control for the possibility that standard models (e.g., Fama-French) may not be correctly specified.
Sorting Process
Sort all stock returns into 10 deciles based on size (market capitalization).
Within each size decile, sort returns into 10 deciles based on B/M ratio.
Calculate the average return of each portfolio for each period (i.e., the expected return for stocks given their characteristics).
Compare each stock’s return to its corresponding portfolio.
Important Notes:
Sorting often leads to more conservative estimates compared to Fama-French models.
If the event study results change depending on the sorting order (e.g., sorting by B/M first vs. size first), this suggests that the findings are not robust.
33.14.2 erer Package
The erer package provides a straightforward implementation of event studies.
Step 1: Load Required Libraries
Step 2: Load Sample Data
The package includes an example dataset, daEsa, which contains stock returns and event dates.
data(daEsa)
head(daEsa)
#> date tb3m sp500 bbc bow csk gp ip kmb
#> 1 19900102 0.3973510 1.76420 2.5352 1.3575 0.6289 4.1237 1.3274 1.8707
#> 2 19900103 0.6596306 -0.25889 0.2747 0.8929 6.2500 0.9901 -0.2183 -0.3339
#> 3 19900104 -0.5242464 -0.86503 -1.3699 -0.4425 -2.3529 0.7353 -0.4376 -0.1675
#> 4 19900105 -0.6587615 -0.98041 -0.5556 -0.4444 1.2048 0.0000 0.2198 -0.6711
#> 5 19900108 0.0000000 0.45043 -1.3966 -0.8929 -1.1905 0.4866 -0.2193 1.0135
#> 6 19900109 0.1326260 -1.18567 0.2833 -0.4505 -2.4096 -0.2421 -2.1978 -2.1739
#> lpx mwv pch pcl pop tin wpp wy
#> 1 1.7341 1.6529 4.0816 1.5464 2.43525 -1.0791 2.9197 2.7149
#> 2 0.8523 2.0325 0.0000 0.5076 1.41509 -1.4545 0.7092 -2.2026
#> 3 -0.2817 0.3984 0.3268 -0.5051 -0.93023 -0.1845 2.1127 -0.9009
#> 4 -0.8475 -0.3968 -0.6515 -0.5076 0.00000 0.5545 -0.6897 -0.4545
#> 5 -0.5698 -0.3984 0.3279 1.0204 -0.93897 -0.5515 -0.6944 0.0000
#> 6 -0.2865 -1.6000 0.3268 -2.5253 -3.79147 -2.4030 0.6993 -1.8265Step 3: Compute Abnormal Returns
We define the estimation window (250 days before the event) and the event window (\(\pm5\) days around the event):
hh <- evReturn(
y = daEsa,
firm = "wpp",
y.date = "date",
index = "sp500",
est.win = 250,
event.date = 19990505,
event.win = 5
)Step 4: Visualizing the Results
Figure 33.6 plots the abnormal-return path produced by erer::evReturn() for the WPP/SP500 example over the \(\pm5\)-day event window.
plot(hh)
Figure 33.6: Abnormal-return path for WPP relative to the S&P 500 over a \(\pm5\)-day event window, generated by the erer package’s evReturn() routine.
33.14.3 Eventus
2 types of output:
-
Using different estimation methods (e.g., market model to calendar-time approach)
Does not include event-specific returns. Hence, no regression later to determine variables that can affect abnormal stock returns.
-
Cross-sectional Analysis of Eventus: Event-specific abnormal returns (using monthly or daily data) for cross-sectional analysis (under Cross-Sectional Analysis section)
- Since it has the stock-specific abnormal returns, we can do regression on CARs later. But it only gives market-adjusted model. However, according to (A. Sorescu, Warren, and Ertekin 2017), they advocate for the use of market-adjusted model for the short-term only, and reserve the FF4 for the longer-term event studies using monthly or daily data.
33.14.3.1 Basic Event Study
- Input a text file containing a firm identifier (e.g., PERMNO, CUSIP) and the event date
- Choose market indices: equally weighted and the value weighted index (i.e., weighted by their market capitalization). And check Fama-French and Carhart factors.
- Estimation options
Estimation period:
ESTLEN = 100is the convention so that the estimation is not impacted by outliers.Use “autodate” options: the first trading day after the event date is used if the event falls on a weekend or holiday
- Abnormal returns window: depends on the specific event
- Choose test: either parametric (including Patell Standardized Residual (PSR)) or non-parametric
33.14.3.2 Cross-sectional Analysis of Eventus
Similar to the Basic Event Study, but now you can have event-specific abnormal returns.
33.15 Practical Guidance on Event Studies
Whether an event study is an appropriate tool depends less on the method itself and more on the setting in which it is deployed. The classical finance setting, a single firm, a precisely dated announcement, liquid continuously priced securities, and an event window short enough that confounding news is unlikely, is where the method earns its reputation. Regulatory rulings whose timing is exogenous, unexpected earnings surprises, sudden executive departures, and M&A announcements all fit. Outside this zone, the method can still be useful, but it begins to inherit the identification challenges of whatever surrounding design is doing the causal work.
The biggest threat to a clean event study is not the statistics but the narrative. Confounding news is the number-one reason event-study estimates are disputed in practice. If a regulatory announcement coincides with an earnings release, a macroeconomic shock, or a peer firm’s news event, the abnormal returns reflect all of it. The remedy is boring but essential: check a news feed for every day in the event window, document what you find, and either pre-commit to a specification that excludes contaminated events or report robustness to their exclusion.
Two inference issues deserve specific attention. Cross-sectional correlation arises when many firms experience the same event on the same day, a sector-wide regulation, a macro shock, a regulatory deadline. Abnormal returns across those firms are not independent, and conventional \(t\)-tests overstate significance. Calendar-time abnormal returns (CTARs) or clustering at the event-date level fix this. Long-horizon studies (BHARs over one to five years) introduce a different problem: new-listing bias, rebalancing bias, and outcome skewness make long-run estimates notoriously fragile, as documented by Barber and Lyon (1997) and Lyon, Barber, and Tsai (1999). Treat long-horizon event studies with caution, and always compare BHAR to calendar-time-portfolio results.
The framing matters as much as the mechanics. As emphasized in the opening note to this chapter, abnormal returns describe market reactions, they are not automatically causal effects of the event on firm value. Claiming causality requires the event to be unanticipated, precisely dated, and exclusive of confounding news. Where those conditions are plausible (a surprise SEC enforcement action, a court ruling whose timing is outside the firm’s control), the event study does identify a causal quantity. Where they are not, the honest framing is “the stock moved around this event by \(X\)%” rather than “the event caused an \(X\)% change in firm value.”
For a complete writeup: justify the event and estimation windows; specify the expected-return model and show its fit in the estimation window; report CAR or ACAR with an appropriate test statistic (Patell, BMP, or a non-parametric rank test, depending on the return distribution); run placebo event studies on non-event dates; and include a narrative audit of the event window for confounding news.
33.16 Econometric Event-Study Designs
The preceding sections of this chapter cover the finance event study: a single information event, a market or factor model fit on a pre-event estimation window, and abnormal returns aggregated over a short event window. That object is descriptive by construction and becomes causal only under the strong identifying assumptions of the [Efficient Market Hypothesis], precise event dating, and absence of confounding news.
A second, distinct object also travels under the name “event study” in modern empirical economics: the dynamic difference-in-differences regression that traces leads and lags around the timing of a treatment in panel data. The two methods share a visual vocabulary, a coefficient indexed by event time plotted against zero, but they differ fundamentally in the data, the estimand, and the identifying assumptions. Because this book sits in the Quasi-Experimental Methods part, it is important that readers leaving this chapter understand both meanings of the term and know where to go for the modern panel design. This section bridges the two.
33.16.1 Finance versus Panel Econometric Event Studies
| Dimension | Finance event study | Econometric event study |
|---|---|---|
| Data structure | Daily returns on a small number of firms; short window | Panel of units over many periods; long pre- and post-treatment |
| Outcome | Abnormal stock return relative to a market or factor model | Any panel outcome (employment, prices, sales, health) |
| Estimand | Cumulative abnormal return \(\text{CAR}\) around an event | Dynamic average treatment effect \(\beta_k\) at event time \(k\) |
| Identifying assumption | Efficient markets, no confounding news, correct return model | Parallel trends, no anticipation, stable composition (panel DiD) |
| Counterfactual | Predicted return from market/factor model | Untreated or not-yet-treated units in the same period |
| Treatment timing | Single calendar date for all firms | Staggered: treatment date \(g_i\) varies across units |
| Inference target | Did the market price the event in? | What is the path of the treatment effect over time? |
Table 33.12 lays out the contrast across the dimensions that most shape interpretation: data structure, outcome, target estimand, identifying assumption, counterfactual, treatment timing, and the question being asked.
The two designs answer different questions. The finance event study asks: did the asset price move when this information arrived, in a way the market model could not have predicted? The econometric event study asks: across many units treated at potentially different dates, what is the average dynamic path of an outcome before and after treatment, relative to comparable untreated units? The first inherits its causal interpretation from the [Efficient Market Hypothesis]; the second inherits it from a Difference-in-Differences design.
33.16.2 The Standard Event-Study Regression
Let \(Y_{it}\) be the outcome for unit \(i\) in period \(t\), and let \(g_i\) denote the period in which unit \(i\) first receives treatment. Define event time as \(k_{it} = t - g_i\), the number of periods since (or until) treatment for unit \(i\). The standard two-way fixed-effects event-study specification is
\[ Y_{it} = \alpha_i + \lambda_t + \sum_{\substack{k = -K \\ k \neq -1}}^{L} \beta_k \cdot \mathbb{1}\{t - g_i = k\} + \varepsilon_{it}, \]
where \(\alpha_i\) are unit fixed effects, \(\lambda_t\) are period fixed effects, and the indicator \(\mathbb{1}\{t - g_i = k\}\) equals one when unit \(i\) is exactly \(k\) periods from its treatment date in period \(t\). The period \(k = -1\) is omitted as the reference category, so every \(\beta_k\) is interpreted as a difference relative to the last pre-treatment period.
Three modeling choices deserve emphasis.
Reference period. Dropping \(k = -1\) is conventional because the last pre-treatment period is typically the cleanest comparison: the unit is not yet treated, but is closest in time to the treated state. Coefficients \(\beta_k\) for \(k < -1\) then function as pre-trend tests. If parallel trends holds, these coefficients should be statistically indistinguishable from zero. Coefficients \(\beta_k\) for \(k \geq 0\) are the dynamic treatment effects: \(\beta_0\) is the on-impact effect, \(\beta_1\) the one-period-after effect, and so on.
Endpoint binning. With unbalanced panels and staggered treatment, some event times are observed for only a handful of units, especially at the extreme leads and lags. Two problems follow. First, \(\beta_k\) at the endpoints is identified from very few observations and therefore noisy. Second, and more subtly, leaving the endpoints unbinned means the regression implicitly extrapolates the treatment effect outside the support of the data, which (Schmidheiny and Siegloch 2023) show can introduce bias in distributed-lag and event-study TWFE models. The standard fix is to bin the endpoints: replace \(\mathbb{1}\{t - g_i = -K\}\) with \(\mathbb{1}\{t - g_i \leq -K\}\) and \(\mathbb{1}\{t - g_i = L\}\) with \(\mathbb{1}\{t - g_i \geq L\}\), so that all events outside the window are absorbed into the boundary indicators. Schmidheiny and Siegloch (2023) establish the equivalence of binned event-study and distributed-lag specifications and provide guidance on choosing \(K\) and \(L\).
Never-treated and not-yet-treated units. Units that never receive treatment have undefined event time; they enter the regression only through the unit and period fixed effects, serving as controls. Some implementations encode their event time as NA, others as a large value such as 10000, which is then excluded from the dummy expansion by the binning rule. The distinction matters in software but not in principle.
33.16.3 Worked Example with fixest
The fixest package implements the event-study regression compactly through the i() operator, which expands a categorical variable into indicators with a chosen reference level. Fitting an event study with leads and lags around treatment timing reduces to one line.
Prose reference: Fig. 33.7 plots the dynamic coefficients from a TWFE event-study regression on simulated staggered-adoption data.
library(fixest)
library(data.table)
# Simulate a staggered-adoption panel.
set.seed(20260430)
n_units <- 200
n_years <- 20
years <- 2001:(2000 + n_years)
# Treatment cohorts: 2008, 2012, 2016, plus a never-treated group.
cohort_pool <- c(2008, 2012, 2016, NA_integer_)
units <- data.table(
id = 1:n_units,
g = sample(cohort_pool, n_units, replace = TRUE)
)
panel <- CJ(id = units$id, year = years)
panel <- merge(panel, units, by = "id")
# Event time; never-treated coded as a large value so it is binned out.
panel[, time_to_treat := ifelse(is.na(g), 1000L, year - g)]
panel[, treated_post := !is.na(g) & year >= g]
# Outcome: unit and year FEs plus a true dynamic effect that grows then plateaus.
true_effect <- function(k) ifelse(k < 0, 0, pmin(0.5 + 0.25 * k, 2.0))
panel[, unit_fe := rnorm(n_units)[id]]
panel[, year_fe := (year - 2000) * 0.05]
panel[, y := unit_fe + year_fe +
ifelse(treated_post, true_effect(time_to_treat), 0) +
rnorm(.N, sd = 0.5)]
# Bin the endpoints at -5 and +8 so the extremes don't drive the estimates.
panel[, ttt_bin := pmin(pmax(time_to_treat, -5L), 8L)]
panel[is.na(g), ttt_bin := -1L] # never-treated absorbed into reference
es_fit <- feols(
y ~ i(ttt_bin, ref = -1) | id + year,
data = panel,
cluster = ~ id
)
iplot(
es_fit,
main = "",
xlab = "Event time (years from treatment)",
ylab = "Estimated coefficient",
ref.line = -1,
pt.join = TRUE
)
Figure 33.7: Event-study coefficients from a two-way fixed-effects regression on simulated staggered-adoption data, with the period before treatment as the reference.
Two implementation details are worth flagging. First, i(ttt_bin, ref = -1) produces one indicator per unique value of ttt_bin other than \(-1\), automatically following the convention that the period immediately before treatment is the reference. Second, clustering at the unit level (cluster = ~ id) handles the within-unit serial correlation that is endemic to panel event studies. Plot output can equivalently be produced through ggiplot::ggiplot(es_fit) for a ggplot2-based version that is easier to customize.
33.16.4 Bias from Heterogeneous Treatment Effects
The TWFE event study above shares a defect with the static Two-Way Fixed Effects DiD estimator: when treatment effects vary across cohorts or across event time, the OLS estimate of any \(\beta_k\) is a weighted average of cohort-specific effects whose weights can be negative, producing a coefficient that does not correspond to any economically meaningful average treatment effect. The mechanism is the same one analyzed at length in Chapter 30.8: already-treated units serve as controls for not-yet-treated units, so heterogeneous post-treatment effects contaminate the comparison. The pre-trend coefficients are similarly polluted, which means a flat pre-trend is not sufficient evidence of parallel trends in a staggered setting.
The cleanest fix that stays within the fixest workflow is the L. Sun and Abraham (2021) estimator, which interacts cohort indicators with event-time indicators and reports a properly weighted average. In fixest, this is one function call.
sa_fit <- feols(
y ~ sunab(g, year) | id + year,
data = panel[!is.na(g)],
cluster = ~ id
)
# iplot(sa_fit) plots the cohort-aggregated dynamic effects.The sunab(g, year) syntax expects the cohort variable g (calendar period of first treatment, with NA reserved for never-treated controls handled separately) and the calendar period year. The estimator returns event-time coefficients \(\beta_k\) that are interaction-weighted across cohorts and immune to the heterogeneous-effects contamination of vanilla TWFE. Other modern alternatives include Callaway and Sant’Anna (2021) (group-time ATTs) and Borusyak, Jaravel, and Spiess (2024) (imputation-based event studies); both are surveyed in Chapter 30.8.
33.16.5 Cross-Reference and Scope
This section is a bridge, not a complete treatment. The goal is to make explicit the conceptual jump from finance event studies (where the counterfactual is a market-model prediction) to econometric event studies (where the counterfactual is an untreated unit), and to flag the standard fixest workflow that most empirical readers will encounter. For full coverage of the modern staggered-adoption estimators, including Callaway and Sant’Anna (2021), L. Sun and Abraham (2021), Borusyak, Jaravel, and Spiess (2024), and the underlying decomposition results that motivate them, readers should consult Chapter 30.8. For the foundational identification framework of parallel trends, no anticipation, and stable composition, see Chapter 30.
The practical upshot for an applied reader: if your data are daily returns around an information event, use the finance machinery in the earlier sections of this chapter. If your data are a panel of units treated at different dates with an outcome you wish to track dynamically, use a DiD-style event study, bin the endpoints, treat the pre-trend coefficients as a diagnostic rather than a proof, and reach for a heterogeneity-robust estimator such as sunab() whenever staggered adoption is involved.
